Next Article in Journal
DICOMization of Proprietary Files Obtained from Confocal, Whole-Slide, and FIB-SEM Microscope Scanners
Previous Article in Journal
Low-Cost, Real-Time Polymerase Chain Reaction System for Point-of-Care Medical Diagnosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SHEL5K: An Extended Dataset and Benchmarking for Safety Helmet Detection

1
Department of Computer Science and Software Engineering, College of Information Technology, United Arab Emirates University, Al Ain 15551, United Arab Emirates
2
RIKEN Center for Brain Science (CBS), Wako 463-0003, Japan
3
Department of Electrical Engineering, National Taipei University of Technology, Taipei 10608, Taiwan
4
College of Artificial Intelligence and Green Energy, National Yang Ming Chiao Tung University, Hsinchu 30010, Taiwan
5
Department of Computer Science, National Yang Ming Chiao Tung University, Hsinchu 30010, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(6), 2315; https://doi.org/10.3390/s22062315
Submission received: 6 January 2022 / Revised: 23 February 2022 / Accepted: 8 March 2022 / Published: 17 March 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Wearing a safety helmet is important in construction and manufacturing industrial activities to avoid unpleasant situations. This safety compliance can be ensured by developing an automatic helmet detection system using various computer vision and deep learning approaches. Developing a deep-learning-based helmet detection model usually requires an enormous amount of training data. However, there are very few public safety helmet datasets available in the literature, in which most of them are not entirely labeled, and the labeled one contains fewer classes. This paper presents the Safety HELmet dataset with 5K images (SHEL5K) dataset, an enhanced version of the SHD dataset. The proposed dataset consists of six completely labeled classes (helmet, head, head with helmet, person with helmet, person without helmet, and face). The proposed dataset was tested on multiple state-of-the-art object detection models, i.e., YOLOv3 (YOLOv3, YOLOv3-tiny, and YOLOv3-SPP), YOLOv4 (YOLOv4 and YOLOv4pacsp-x-mish), YOLOv5-P5 (YOLOv5s, YOLOv5m, and YOLOv5x), the Faster Region-based Convolutional Neural Network (Faster-RCNN) with the Inception V2 architecture, and YOLOR. The experimental results from the various models on the proposed dataset were compared and showed improvement in the mean Average Precision (mAP). The SHEL5K dataset had an advantage over other safety helmet datasets as it contains fewer images with better labels and more classes, making helmet detection more accurate.

1. Introduction

Workplace safety has become a focus for many production and work sites due to the consequences of the unsafe environment on the health and productivity of the workforce. According to statistics [1,2,3,4], the construction industry is at high risk for the injuries and deaths of workers. In 2005, the National Institute for Occupational Safety and Health (NIOSH) reported 1224 deaths of construction workers in 1 y, making it the most dangerous industry in the United States (U.S.) [1]. Moreover, the U.S. Bureau of Labor Statistics (BLS) estimated injuries for 150,000 workers every year at construction sites [1]. The Bureau also reported the death of one in five workers in 2014 and a total of 1061 construction workers’ deaths in 2019 [2,3]. As per the report of the Ministry of Employment and Labor (MEOL) in Korea, 964 and 971 workers died in workplace accidents in 2016 and 2017, respectively [4]. Among these fatalities, 485 fatalities occurred at construction sites, followed by 217 and 154 in the manufacturing and service industry, respectively. Workers at most of the worksites and manual working environments are at high risk of injuries because of not following the safety measures and using Personal Protective Equipment (PPE). The carelessness of the workers and not following PPE compliance will have adverse effects and pose more threats of minor or major injuries. In 2012, the National Safety Council (NSC) reported more than 65,000 cases of head injuries and 1020 deaths at construction sites [5]. According to the American Journal of Industrial Medicine, a total number of 2210 construction workers died because of a Traumatic Brain Injury (TBI) from 2003 to 2010 [6]. Released by Headway, the brain injury association, 3% of PPE purchased was for head protection as head injuries account for more than 20% of total injuries [7].
These statistics delineate the prevalence of fatal and non-fatal injuries in the construction industry, and there is a dire need to reduce the rate. Creating a safe environment for workers brings an arduous challenge for this sector globally. Adopting safety measures and providing construction workers with PPE can result in decreasing accident rates. Despite the effectiveness of these strategies, it is not guaranteed that the workers would be cautious and use the PPE. To avert all these troubles, there is a need to discover automated ways of detection and monitoring safety helmets. A deep-learning-based safety helmet detection system can be developed by using a large amount of labeled data. However, there is a lack of datasets to build highly accurate deep learning models for workers’ helmet detection. There are few publicly available datasets for safety helmet detection, which are not entirely labeled, and the labeled ones contain fewer classes and incomplete labels. Therefore, the proposed work presents the Safety HELmet dataset with 5K images (SHEL5K) dataset, an enhanced version of the SHD dataset [8]. In the SHD dataset [8], many objects are not labeled, which is not sufficient to train an efficient helmet recognition model. The SHD dataset [8] was improved in the proposed work by labeling all three originally proposed classes and adding three more classes for training an efficient helmet detection model. The main aims of the proposed study were to: (1) complete the missing labels and (2) increase the number of classes from three to six (helmet, head with helmet, person with helmet, head, person without helmet, and face). The proposed dataset was tested on various object detection models, i.e., YOLOv3 [9], YOLOv3-tiny [10], YOLOv3-SPP [11], YOLOv4 [12], YOLOv5-P5 [13], the Faster Region-based Convolutional Neural Network (Faster-RCNN) [14] with the Inception V2 architecture [15], and YOLOR [16] models. The experimental results showed significant improvements in the mAP as compared to the publicly available datasets. A comparative analysis was performed, and discussions are provided based on results from the various models. The proposed system was also used to successfully perform real-time safety helmet detection in YouTube videos.

2. Related Work

In the literature, various efforts have been made by researchers to develop a vision-based system for the helmet detection task. Li et al. [17] proposed a Convolutional-Neural-Network (CNN)-based safety helmet detection method using a dataset of 3500 images collected by the web crawling method. The precision and recall of the system were recorded as 95% and 77%, respectively. Wang et al. [18] proposed a safety helmet detection model trained on a total of 10,000 images captured by 10 different surveillance cameras at construction sites. In the experiment’s first phase, the authors employed the YOLOv3 architecture [9] and achieved an mAP0.5 of 42.5%. In the second phase, the authors improved the architecture of YOLOv3 [18] and achieved an mAP0.5 of 67.05%. Wang et al. [19] suggested a hardhat detection system based on a lightweight CNN using the Harvard database hardhat dataset [20]. The dataset contains 7064 annotated images, which consist of three classes (helmet, head, and person). In the three classes, the person class is not appropriately labeled. The network was trained considering two classes (helmet and head) and achieved an average accuracy of 87.4% and 89.4% for head and helmet, respectively. Li et al. [21] trained an automatic safety helmet-wearing detection system using the INRIA person dataset [22] and collected pedestrian data from a power substation. The authors in [21] showed that the accuracy of the proposed method (Color Feature Discrimination (CFD) and the ViBE algorithm in combination with the c4 classifier) yielded better results than HOG features and the SVM classifier method. The accuracy of the HOG feature with the SVM classifier achieved 89.2%, while the proposed method achieved an accuracy of 94.13%. Rubaiyat et al. [23] proposed an automated system for detecting helmets in construction safety. The authors collected 1000 images from the Internet using a web crawler, which consisted of 354 human images and 600 non-human images. The helmet class achieved an accuracy of 79.10%, while the without helmet class achieved an accuracy of 84.34%. Similarly, Kamboj and Powar [24] proposed an efficient deep-learning-based safety helmet detection system for the industrial environment by acquiring data from various videos of an industrial facility. The videos were captured by using cameras having a resolution of 1920 × 1080 px and a frame rate of 25 frames per second. The dataset consisted of 5773 images having two classes (helmet and without helmet). An improved helmet detection was proposed by Geng et al. [25] using an imbalanced dataset of 7581 images, mostly with a person in a helmet and a complex background. The label confidence of 0.982 was achieved by testing it on 689 images. Moreover, Long et al. [26] proposed a deep-learning-based detection of safety helmet wearing using 5229 images, acquired from the Internet and various power plants (including power plants under construction). The proposed system was based on SSD, and an mAP0.5 of 78.3% was achieved on the test images and compared with SSD, which was 70.8% using an IoU of 0.5. In the above studies [17,18,23,24,26], they used custom data to test their method; therefore, it is not fair to make a comparison of the proposed work in this paper with these methods.

2.1. Datasets for Safety Helmet Detection

In general, researchers develop helmet detection systems using custom data or publicly available datasets. Some of the publicly available datasets, i.e., [8,20,27,28], for safety helmet detection are summarized in Table 1. Table 1 shows a brief comparison of the proposed dataset in the current study with various publicly available datasets. Each dataset shown in Table 1 is explained in detail below.

2.1.1. Safety Helmet Detection Dataset

The Safety Helmet Detection (SHD) dataset [8] is a publicly available dataset on Kaggle containing 5000 labeled images and three classes (helmet—18,966, head—5785, and person—751). However, the dataset has many incompletely labeled objects. Figure 1b shows the dataset labels, which shows that the person class is not labeled.

2.1.2. Hardhat Dataset

The hardhat dataset [20] is a safety helmet dataset shared by Northeastern University consisting of 7063 labeled images. The dataset is divided into training and testing sets, which contain 5297 and 1766 images, respectively. The images are from three distinct classes having 27,249 labeled objects (helmet—19,852, head—6781, and person—616). In the given dataset, the person class is not labeled properly, as shown in Figure 1c, and the number of images in each class is not distributed equally.

2.1.3. Hard Hat Workers Object Detection Dataset

The Hard Hat Workers (HHW) dataset [27] is an improved version of the hardhat dataset [20] and is publicly available on the Roboflow website. In the HHW dataset [27], the number of labels in each class is increased (helmet—26,506, head—8263, and person—998). Figure 1d shows a sample image of the HHW dataset [27] labels in which it can be seen that the person class is not labeled.

2.1.4. Safety Helmet Wearing Dataset

The Safety Helmet Wearing (SHW) dataset [28] consists of 7581 images. The images have 111,514 safety helmet-wearing or positive class objects and 9044 not-wearing or negative class objects. Some of the negative class objects were obtained from the SCUT-HEAD dataset [29]. Several bugs of the original SCUT-HEAD dataset [29] were fixed to directly load the data into a normal PASCAL VOC format. Most images in the dataset are helmet images, and there are a very small number of head images. Figure 1e shows a labeled sample image from the SHW dataset. Figure 1a shows a comparison between the public datasets’ labels and the SHEL5K dataset’s labels.

3. SHEL5K Dataset

In the proposed work, the number of labels and classes in the SHD dataset [8] were extended and completed. Figure 2 shows sample images of the SHD dataset [8]. The SHD dataset [8] contains 5000 images having a resolution of 416 × 416 and 25,501 labels with complicated backgrounds and bounding box annotations in PASCAL VOC format for the three classes namely helmet, head, and person. The limitation of the SHD dataset [8] is that numerous objects are incompletely labeled. Figure 3a,b shows image samples with person and head not properly labeled. The main aims of the proposed study were to: (1) completed the missing labels and (2) increase the number of classes from three to six (helmet, head with helmet, person with helmet, head, person without helmet, and face).
To address the limitations associated with the SHD dataset, SHEL5K is proposed, which consists of 75,570 labels. The number of labels in the SHEL5K dataset was increased for each class, i.e., (helmet—19,252, head—6120, head with helmet—16,048, person without helmet—5248, person with helmet—14,767, and face—14,135). Figure 3 shows the comparison of the labels of the SHD dataset [8] (a and b) and SHEL5K datasets (c and d), with the helmet in blue, the head in purple, the head with helmet in navy blue, the person with helmet in green, the person without a helmet in red, and the face in the yellow bounding boxes. Moreover, the graph in Figure 4 shows the comparison of the SHD dataset [8] and SHEL5K dataset in terms of the number of labels of each class. The SHD dataset [8] and SHEL5K labels are represented by blue and orange bars, respectively. From the graph, it can be seen that the class person is too poorly labeled. In the proposed work, the labeling of the image was performed by using the LabelImg [30] tool with the following steps: (1) the default number of classes in the tool was changed to six for our dataset; (2) images opening and label saving paths were specified; (3) objects corresponding to the classes were labeled, and an XML file was created.
The file contains the name of the image, the path to the image, the image size and depth, and the coordinates of the producer image.

4. Results and Discussion

The proposed dataset SHEL5K was benchmarked by using state-of-the-art one-stage object detection models such as YOLOv3 [9], YOLOv4 [12] YOLOv5-P5 [13], the Faster-RCNN [14] with Inception v2 [15], and YOLOR [16]. In particular, we employed different pretrained variations of the models, i.e., YOLOv3-tiny [10], YOLOv3 [9], YOLOv3-SPP [11], YOLOv3-SPP pretrained on the MS COCO dataset [31], YOLOv3-SPP pretrained on the ImageNet dataset [32], and YOLOv5-P5 models (YOLOv5s, YOLOv5m, YOLOv5x) [13]. These models were prepared using the COCO 128 dataset, which contains the first 128 images of COCO train 2017 [31].

4.1. Evaluation Metrics

In the proposed work, the precision, recall, F1 score, and mAP were used as the evaluation metrics to perform a fair comparison between the experimental results of the models. The precision represents the object detection model’s probability of the predicted bounding boxes being identical to the actual ground truth boxes and is described in Equation (1) below.
Precision = T P ( T P + F P )
where TP, TN, FP, and FN refer to True Positive, True Negative, False Positive, and False Negative, respectively. The recall represents the probability of ground truth objects being correctly detected as depicted in (2).
Recall = T P ( T P + F N )
Moreover, the F1 score is the harmonic mean of the model’s precision and recall, and the mathematical representation is shown in Equation (3).
F 1 score = 2 Precision Recall Precision + Recall
Additionally, the mean Average Precision (mAP) is the score achieved by comparing the detected bounding box to the ground truth bounding box. If the intersection over union score of both the boxes is 50% or larger, the detection is considered as TP. The mathematical formula of the mAP is given in Equation (4) below.
m A P = 1 n k = 1 k = n A P k
where A P k is the average precision of class k and n represents the number of classes.

4.2. Experimental Setup

Data preparation started with the conversion of annotated files from the PASCAL VOC format to the YOLO format to be given as the input to object detection models. The proposed dataset was randomly divided into training and testing sets. The training set contained a total of 4000 (80%) images, while the testing set contained 1000 (20%) images. The criterion for evaluating the performance of the various models were the mAP0.5 (the and F1 score. During the experiments, the Intersection over Union (IoU) threshold value was kept at 0.5. YOLOv3-SPP [11], YOLOv4 [12], YOLOv5-P5 [13], and YOLOR [16] were considered trained on the proposed dataset as these models have the fastest inference time for real-time object detection as compared to the majority of object detection models. The reason is that these models perform classification and bounding box regression in a single step. Empirically, it was found that the suitable number of epochs for training the YOLOv3 models (YOLOv3-tiny [10], YOLOv3 [9], and YOLOv3-SPP [11]), and the Faster-RCNN [14] with Inception v2 [15] was 1000, while for the other models, namely YOLOv4 [12], YOLOv5-P5 [13], and YOLOR [16], the value was 500. The performance of these models was also compared with the Faster-RCNN [14] with Inception v2 [15] model, which is better at detecting small objects. The Faster-RCNN [14] with Inception V2 [15] model was trained with 250,000 steps, and the learning rate was set to 0.0002 while keeping the value of the batch size equal to 16. The results of the Faster-RCNN [14] with Inception v2 [15] were measured using the comparative analysis of object detection metrics with a companion open-source toolkit [33], which is similar to the YOLO models.

4.3. Three-Class Results

The SHD dataset [8] has three classes (helmet, head, and person) and the SHEL5K dataset has six classes (helmet, head with a helmet, a person with a helmet, head, person without a helmet, and face). In the current study, two classes person with helmet and person without helmet were combined to perform a fair comparison between the two datasets. Both classes were merged, as they correspond to the class person in the SHD dataset [8]. Figure 5 shows the comparison of the SHD and SHEL5K dataset results on the same images. For the SHD dataset [8] results, the person class was not detected. Table 2 shows the comparison results of the YOLOv3-SPP [11] and YOLOv5x [13] models on the SHD dataset [8] and SHEL5K dataset. For the sake of simplicity, the results of two models, YOLOv3-SPP [11] and YOLOv5x [13], are presented as they outperformed the remaining models. For the YOLOv5x [13] model, an mAP0.5 of 0.8528 was achieved for three classes, where the best and worst mAP0.5s were 0.8774 and 0.8311 for the helmet and person classes, respectively. The trained model showed low performance in the case of the person class in comparison with the helmet class. The YOLOv5x [13] model achieved better performance than YOLOv3-SPP [11] as shown in Table 2. The head class in the SHD dataset [8] achieved high a precision, recall and F1 score as it was properly labeled in comparison with the other classes. The helmet class results did not perform well as head with helmet and helmet were given a single label helmet in the dataset. Moreover, the results of the person class were low as the labeling of the person class was incomplete in the SHD dataset [8].
Figure 6 shows the confusion matrices for the YOLOv5x [13] model based on various publicly available datasets. The confusion matrices computed for the HHW dataset [27], the hardhat dataset [20], and the SHD dataset [8] for three classes (helmet, head, and person) are plotted in Figure 6a–c, respectively. The confusion matrices showed very poor results for the class person, and the background FNs were also high. The background FNs are the unrecognized percentage of the labeled object, and its results were for all three datasets. The helmet and head classes performed very well, and the background FPs were recorded as high. Figure 6d shows the confusion matrices constructed by the object detection models on the SHW dataset [28] test dataset for two classes (hat and person). The confusion matrix shows that the YOLOv5x [13] model showed good performance on the SHW dataset [28] as compared to the other datasets. Overall, the performance of the other datasets was also promising, except for the person class, which was not detected by the model. Therefore, in the current study, the dataset was extended and labeled properly, and additional classes were added to have a more accurate detection of the person class in the SHEL5K dataset.

4.4. Six-Class Results

Table 3 and Table 4 show the comparison results of different variations of the YOLOv3-SPP [11] and YOLOv5-P5 [13] models trained on the SHEL5K dataset. For YOLOv3-SPP [11], three different models were evaluated using the SHEL5K dataset, in which one was trained from scratch (not pretrained) and the other two models were pretrained on the ImageNet dataset [32] and MS COCO dataset [31]. The highest mAP0.5 of 0.5572 was achieved by the YOLOv3-SPP [11] model pretrained on the ImageNet dataset [32]. For YOLOv3-SPP [11], the highest mAP0.5 of 0.6459 was achieved for the head with helmet class, while the two worst mAP0.5s of 0.007 and 0.0295 were reported for the face class when the model was trained from scratch and YOLOv3-SPP [11] was pretrained on the MS COCO dataset [31]. These models achieved an mAP0.5 value of nearly zero, which may be because the human faces were far away in most images and there was no face class included in the COCO dataset [31]. For the YOLOv5-P5 [13] model, Table 4 shows a comparison of the results of the three models of YOLOv5-P5 [13] on the SHEL5K dataset. The YOLOv5-P5 [13] model is available in the YOLOv5s, YOLOv5m, YOLOv5l, and YOLOv5x models. However, in the current study, the YOLOv5s, YOLOv5m, and YOLOv5x models on the pretrained COCO128 dataset [31] were selected. The YOLOv5x [13] achieved an mAP0.5 of 0.8033, and the class with the highest mAP0.5 of 0.8565 was the class person with helmet. The results of the face class were relatively poor, and the mAP0.5 was 0.7196. The mAP0.5 of YOLOv5-P5 [13] was better than YOLOv3-SPP [11]. The results of the YOLOv5x [13] model on three different types of images captured at various distances (far, near, and medium) are shown in Figure 7.
Figure 8 shows the confusion matrix of the SHEL5K dataset. The results were relatively low compared to the other public datasets. This is also evident from Table 5, which shows the comparison results of the YOLOv5x [13] model on various datasets including the SHEL5K dataset. The model trained on the SHEL5K dataset showed better results compared to the other datasets except for the SHW dataset [28]. The precision, recall, and F1 score achieved by the model on the proposed dataset were slightly lower than the SHW dataset [28]. The precision, recall, and F1 score of the model on the SHEL5K were recorded as 0.9188, 0.817, and 0.8644, respectively. This is because the SHW dataset [28] contains only two classes, while the proposed dataset contains six classes. Moreover, during the labeling of the proposed dataset, an image containing some part of the helmet and face was labeled as the helmet or face class, respectively. The Precision–Recall (PR) curve is also shown in Figure 8, which also depicts that the lowest mAP0.5 (0.72) was achieved by the face class, which was less than the mAP (0.80) of all the other classes.
The results of the YOLOR model on the proposed dataset (SHEL5K) are summarized in Table 6. The YOLOR [16] model used in the proposed work was pretrained on the COCO dataset [31]. The model achieved an mAP0.5 of 0.8828, and the highest mAP0.5 of 0.911 was recorded for the class head with helmet. The result of the class person without helmet was relatively poor with an mAP0.5 of 0.8498. The results of the model on the sample images are depicted in Figure 9.
Figure 10 compares the visualization results of the best model trained on the SHW dataset [28] and the SHEL5K dataset on a test image. It can be seen from the result of the model trained on the SHW dataset [28] in Figure 10a that the model was not able to detect the helmet class if the helmet in the image was half visible and the head of the worker was hidden, as shown in Figure 10a. The results of the model trained on the SHEL5K dataset are shown in Figure 10b, which shows that the model can detect the helmet class correctly, which shows that the labeling in the proposed dataset was performed efficiently. The state-of-the-art model trained on the SHEK5K dataset in the current study did not perform well. However, in the future, the proposed dataset will be given to new object detection models to achieve high performance.
The K-fold cross-validation method was used to check whether the models were subjected to overfitting on the proposed data or not. The proposed dataset was divided into training 80% (4000 images) and testing 20% (1000 images). The value of K was considered five where the data were split into five folds, i.e., K1, K2, K3, K4 and K5. Table 7 shows the results of K-fold cross-validation on the SHEL5K dataset using the YOLOR [16] model. The results of all the folds were comparable, which shows that the model was not subjected to overfitting. The maximum mAP0.5 value of 0.8881 was achieved at fold K5, and the minimum mAP0.5 value of 0.861 was achieved at fold K4.
The results of all the state-of-the-art models trained on the SHEL5K dataset are summarized in Table 8. The performance of the YOLO models was compared with the Faster-RCNN with the Inception V2 architecture. YOLOv3-tiny [10], YOLOv3 [9], and YOLOv3-SPP [11] were the models pretrained on the ImageNet dataset [32], while YOLOv5s, YOLOv5m, and YOLOv5x [13] were pretrained on the COCO128 dataset [31]. Detection results of the best yolov5x [13] models trained on SHEL5K dataset and other publicly available datasets [8,20,27,28] are illustrated in Appendix A. The best mAP0.5 of 0.8828 was achieved by the YOLOR [16] model with a precision, recall, and F1 score of 0.9322, 0.8066, and 0.8637, respectively. The lowest mAP0.5 score of 0.3689 was achieved by the Faster-RCNN [14] with a precision, recall, and F1 score of 0.7808, 0.3862, and 0.5167, respectively. The Faster-RCNN model achieved the highest inference time of 0.05 s. In the YOLO models, the YOLOv3-tiny [10] achieved the lowest mAP0.5 score of 0.3779 with a precision, recall, and F1 score of 0.7695, 0.4225, and 0.5408, respectively. Table 8 show the training time and testing time of all the models. The YOLOv3 tiny model had the lowest inference time of 0.006 s and fewer layers and parameters as compared to the other YOLO models. The YOLOR model achieved the highest mAP0.5 of 0.8828 with an optimum inference time of 0.012 s.

5. Conclusions

The proposed work aimed to extend the number of classes and labels of the publicly available SHD dataset [8]. The SHD dataset [8] contains 5000 images with three object classes (helmet, head, and person); however, most of the images were incompletely labeled. Therefore, a new dataset named SHEL5K (publicly available at https://data.mendeley.com/datasets/9rcv8mm682/draft?a=28c11744-48e7-4810-955b-d76e853beae5 (accessed on 5 January 2022)) was proposed by adding three more classes and completely labeling all 5000 images of the SHD dataset [8]. The proposed dataset was benchmarked on the various state-of-the-art one-stage object detection models, namely YOLOv3-tiny [10], YOLOv3 [9], YOLOv3-SPP [11], YOLOv4 [12], YOLOv5-P5 [13], the Faster-RCNN [14] with Inception v2 [15], and YOLOR [16]. The experimental results showed significant improvements in the mAP0.5 s of the compared models. From the experimental result of the models on the proposed dataset (SHEL5K), it can be concluded that all the models showed promising performances in detecting all classes. It can also be concluded that the proposed dataset had an advantage over the SHD dataset [8] in terms of images and labeling. Moreover, models trained on the proposed dataset can be used for a real-time safety helmet detection task. In the future, we will improve the real-time recognition rate of the safety helmet detection focusing on misclassified cases.

Author Contributions

Software, M.-E.O.; Supervision, M.G.; Writing—original draft, M.-E.O. and M.G.; Writing—review & editing, M.-E.O., M.G., F.A., L.A., T.-H.T., J.-W.H. and P.-Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by United Arab Emirates University. (grant no. 12R012).

Institutional Review Board Statement

Not available.

Informed Consent Statement

We used a publicly available dataset and do not need to be informed.

Data Availability Statement

The dataset is publicly available in Mendeley Data and can be found at: https://data.mendeley.com/datasets/9rcv8mm682/draft?a=28c11744-48e7-4810-955b-d76e853beae5 (accessed on 5 January 2022).

Acknowledgments

The authors would like to thank everyone who helped manually label the dataset and UAEU for providing the DGX-1 supercomputer.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Predictions of the best model trained on the SHEL5k dataset and the ground truth labels of the sample images.
Figure A1. Predictions of the best model trained on the SHEL5k dataset and the ground truth labels of the sample images.
Sensors 22 02315 g0a1
Figure A2. Predictions of the best model trained on the hardhat dataset [20] and the ground truth labels of the sample images.
Figure A2. Predictions of the best model trained on the hardhat dataset [20] and the ground truth labels of the sample images.
Sensors 22 02315 g0a2
Figure A3. Predictions of the best model trained on the SHW dataset [28] and the ground truth labels of the sample images.
Figure A3. Predictions of the best model trained on the SHW dataset [28] and the ground truth labels of the sample images.
Sensors 22 02315 g0a3
Figure A4. Predictions of the best model trained on the SHD dataset [8] and the ground truth labels of the sample images.
Figure A4. Predictions of the best model trained on the SHD dataset [8] and the ground truth labels of the sample images.
Sensors 22 02315 g0a4
Figure A5. Predictions of the best model trained on the HHW dataset [27] and the ground truth labels of the sample images.
Figure A5. Predictions of the best model trained on the HHW dataset [27] and the ground truth labels of the sample images.
Sensors 22 02315 g0a5

References

  1. Keller, J.R. Construction Accident Statistics. Available online: https://www.2keller.com/library/construction-accident-statistics.cfm (accessed on 5 January 2022).
  2. U.S. Bureau of Labor Statistics (BLS). National Census of Fatal Occupational Injuries in 2014. September 2015. Available online: https://www.bls.gov/news.release/archives/cfoi_09172015.pdf (accessed on 5 January 2022).
  3. U.S. Bureau of Labor Statistics (BLS). Census of Fatal Occupational Injuries Summary, 2019. December 2020. Available online: https://www.bls.gov/news.release/cfoi.nr0.html (accessed on 5 January 2022).
  4. Jeon, J.-h. 971 S. Korean Workers Died on the Job in 2018, 7 More than Previous Year. May 2019. Available online: https://www.hani.co.kr/arti/english_edition/e_national/892709.html (accessed on 5 January 2022).
  5. HexArmor. The Hard Truth about Safety Helmet Injuries and Statistics. June 2019. Available online: https://www.hexarmor.com/posts/the-hard-truth-about-safety-helmet-injuries-and-statistics (accessed on 5 January 2022).
  6. Konda, S.; Tiesman, H.M.; Reichard, A.A. Fatal traumatic brain injuries in the construction industry, 2003–2010. Am. J. Ind. Med. 2016, 59, 212–220. [Google Scholar] [CrossRef]
  7. Headway the Brain Injury Association. Workplace hArd Hat Safety Survey Results. Available online: https://www.headway.org.uk/media/8785/workplace-hard-hat-safety-survey-results.pdf (accessed on 5 January 2022).
  8. Larxel. Safety Helmet Detection. Available online: https://www.kaggle.com/andrewmvd/hard-hat-detection (accessed on 5 January 2022).
  9. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  10. Adarsh, P.; Rathi, P.; Kumar, M. YOLO v3-Tiny: Object Detection and Recognition using one stage improved model. In Proceedings of the 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS), Tamil Nadu, India, 6–7 March 2020. [Google Scholar] [CrossRef]
  11. Zhang, X.; Gao, Y.; Wang, H.; Wang, Q. Improve YOLOv3 using dilated spatial pyramid module for multi-scale object detection. Int. J. Adv. Robot. Syst. 2020, 17. [Google Scholar] [CrossRef]
  12. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  13. Jocher, G.; Stoken, A.; Borovec, J.; Liu, C.; Hogan, A. Ultralytics/yolov5: v4.0-nn.SiLU() activations, Weights & Biases logging, PyTorch Hub integration (v4.0). Zenodo 2021. [Google Scholar] [CrossRef]
  14. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R -CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  15. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception architecture for computer vision. arXiv 2015, arXiv:1512.00567. [Google Scholar]
  16. Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. You Only Learn One Representation: Unified Network for Multiple Tasks. arXiv 2021, arXiv:2105.04206. [Google Scholar]
  17. Li, Y.; Wei, H.; Han, Z.; Huang, J.; Wang, W. Deep Learning-Based Safety Helmet Detection in Engineering Management Based on Convolutional Neural Networks. Adv. Civ. Eng. 2020, 2020, 9703560. [Google Scholar] [CrossRef]
  18. Wang, H.; Hu, Z.; Guo, Y.; Yang, Z.; Zhou, F.; Xu, P. A Real-Time Safety Helmet Wearing Detection Approach Based on CSYOLOv3. Appl. Sci. 2020, 10, 6732. [Google Scholar] [CrossRef]
  19. Wang, L.; Xie, L.; Yang, P.; Deng, Q.; Du, S.; Xu, L. Hardhat-Wearing Detection Based on a Lightweight Convolutional Neural Network with Multi-Scale Features and a Top-Down Module. Sensors 2020, 20, 1868. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Xie, L. Hardhat. Harvard Dataverse. 2019. Available online: https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/7CBGOS (accessed on 5 January 2022).
  21. Li, K.; Zhao, X.; Bian, J.; Tan, M. Automatic Safety Helmet Wearing Detection. In Proceedings of the 2017 IEEE 7th Annual International Conference on CYBER Technology in Automation, Control, and Intelligent Systems (CYBER), Honolulu, HI, USA, 31 July–4 August 2017. [Google Scholar] [CrossRef] [Green Version]
  22. Dalal, N. INRIA Person Dataset. Available online: http://pascal.inrialpes.fr/data/human/ (accessed on 5 January 2022).
  23. Rubaiyat, A.H.M.; Toma, T.T.; Kalantari-Khandani, M.; Rahman, S.A.; Chen, L.; Ye, Y.; Pan, C.S. Automatic Detection of Helmet Uses for Construction Safety. In Proceedings of the 2016 IEEE/WIC/ACM International Conference on Web Intelligence Workshops (WIW), Omaha, NE, USA, 13–16 October 2016. [Google Scholar] [CrossRef]
  24. Kamboj, A.; Powar, N. Safety Helmet Detection in Industrial Environment using Deep Learning. In Proceedings of the 9th International Conference on Information Technology Convergence and Services (ITCSE 2020), Zurich, Switzerland, 30–31 May 2020. [Google Scholar] [CrossRef]
  25. Geng, R.; Ma, Y.; Huang, W. An improved helmet detection method for YOLOv3 on an unbalanced dataset. arXiv 2020, arXiv:2011.04214. [Google Scholar]
  26. Long, X.; Cui, W.; Zheng, Z. Safety Helmet Wearing Detection Based On Deep Learning. In Proceedings of the 2019 IEEE 3rd Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chengdu, China, 15–17 March 2019. [Google Scholar] [CrossRef]
  27. Northeastern University—China. Hard Hat Workers Object Detection Dataset. Available online: https://public.roboflow.com/object-detection/hard-hat-workers (accessed on 5 January 2022).
  28. Safety-Helmet-Wearing-Dataset. Available online: https://github.com/njvisionpower/Safety-Helmet-Wearing-Dataset (accessed on 5 January 2022).
  29. Peng, D.; Sun, Z.; Chen, Z.; Cai, Z.; Xie, L.; Jin, L. Detecting Heads using Feature Refine Net and Cascaded Multi-scale Architecture. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–22 August 2018. [Google Scholar] [CrossRef] [Green Version]
  30. Tzutalin. LabelImg. Git Code. 2015. Available online: https://github.com/tzutalin/labelImg (accessed on 5 January 2022).
  31. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. arXiv 2014, arXiv:1405.0312. [Google Scholar]
  32. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet largescale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  33. Padilla, R.; Passos, W.L.; Dias, T.L.B.; Netto, S.L.; da Silva, E.A.B. A Comparative Analysis of Object Detection Metrics with a Companion Open-Source Toolkit. Electronics 2021, 10, 279. [Google Scholar] [CrossRef]
Figure 1. Comparison of public safety helmet datasets’ labels and SHEL5K dataset’s labels: (a) SHEL5K dataset, (b) SHD dataset [8], (c) hardhat dataset [20], (d) HHW dataset [27], and (e) SHW dataset [28].
Figure 1. Comparison of public safety helmet datasets’ labels and SHEL5K dataset’s labels: (a) SHEL5K dataset, (b) SHD dataset [8], (c) hardhat dataset [20], (d) HHW dataset [27], and (e) SHW dataset [28].
Sensors 22 02315 g001
Figure 2. Sample images of the SHEL5K dataset.
Figure 2. Sample images of the SHEL5K dataset.
Sensors 22 02315 g002
Figure 3. (a,b) SHD dataset [8] labels; (c,d) SHEL5K dataset labels.
Figure 3. (a,b) SHD dataset [8] labels; (c,d) SHEL5K dataset labels.
Sensors 22 02315 g003
Figure 4. Bar graph comparison between the SHD dataset [8] and SHEL5K dataset in terms of the number of labels for each class.
Figure 4. Bar graph comparison between the SHD dataset [8] and SHEL5K dataset in terms of the number of labels for each class.
Sensors 22 02315 g004
Figure 5. Comparison of the SHD and SHEL5K dataset results on the same images. (ac) The results of the best SHD dataset [8] model. (df) The results of the best SHEL5K dataset model.
Figure 5. Comparison of the SHD and SHEL5K dataset results on the same images. (ac) The results of the best SHD dataset [8] model. (df) The results of the best SHEL5K dataset model.
Sensors 22 02315 g005
Figure 6. Confusion matrices of the YOLOv5x model on (a) the SHD dataset [8], (b) hardhat dataset [20], (c) HHW dataset [27], and (d) SHW dataset [28].
Figure 6. Confusion matrices of the YOLOv5x model on (a) the SHD dataset [8], (b) hardhat dataset [20], (c) HHW dataset [27], and (d) SHW dataset [28].
Sensors 22 02315 g006
Figure 7. The YOLOv5x [13] detected outputs are plotted with the original images.
Figure 7. The YOLOv5x [13] detected outputs are plotted with the original images.
Sensors 22 02315 g007
Figure 8. Confusion matrix and PR curve of the object detection model calculated on the SHEL5K dataset and the YOLOv5x [13] model.
Figure 8. Confusion matrix and PR curve of the object detection model calculated on the SHEL5K dataset and the YOLOv5x [13] model.
Sensors 22 02315 g008
Figure 9. Result of the YOLOR [16] model experiments on the sample images.
Figure 9. Result of the YOLOR [16] model experiments on the sample images.
Sensors 22 02315 g009
Figure 10. Comparison of the best-trained model results on the (a) SHW dataset [28] and the (b) SHEL5K dataset.
Figure 10. Comparison of the best-trained model results on the (a) SHW dataset [28] and the (b) SHEL5K dataset.
Sensors 22 02315 g010
Table 1. Comparison of public safety helmet datasets and the SHEL5K dataset.
Table 1. Comparison of public safety helmet datasets and the SHEL5K dataset.
DatasetsHardhat [20]HHW [27]SHD [8]SHW [28]SHEL5K
Total sample70637041500075815000
Class33326
Number of labels in each class
Helmet19,85226,50618,966-19,252
Head678182635785-6120
Person *6169987519044-
Head and helmet----16,048
Person not helmet----5248
Person and helmet----14,767
Face----14,135
Hat **---111,514-
Total27,24935,76725,502120,55875,570
* The person class of the SHD dataset is called head. ** The hat class of the SHD dataset is called helmet.
Table 2. Comparison between two dataset results for 3 classes on the YOLOv3-SPP [11] and YOLOv5x [13] models.
Table 2. Comparison between two dataset results for 3 classes on the YOLOv3-SPP [11] and YOLOv5x [13] models.
YOLOv3-SPP [11]
SHD Dataset [8]SHEL5K Dataset with 3 Classes
ClassPrecisionRecallmAP0.5F1PrecisionRecallmAP0.5F1
Helmet0.95780.49760.48690.65490.92220.71970.70280.8084
Head0.91540.3020.29230.45420.91140.66420.64840.7684
Person00000.90920.63540.61480.748
Average0.62440.26650.25970.36970.91430.67310.65530.775
YOLOv5x [13]
SHD Dataset [8]SHEL5K Dataset with 3 Classes
ClassPrecisionRecallmAP0.5F1PrecisionRecallmAP0.5F1
Helmet0.95590.91620.91620.93560.94020.88580.87740.9122
Head0.9090.8790.86860.89380.92160.85620.84990.8877
Person0.03450.00520.00030.0090.92030.84090.83110.8788
Average0.63310.60010.5950.61280.92740.8610.85280.8929
Table 3. Comparison results of different variations of the YOLOv3 models (a) trained from scratch, (b) pretrained on the ImageNet dataset [32], and (c) pretrained on the MS COCO dataset [31].
Table 3. Comparison results of different variations of the YOLOv3 models (a) trained from scratch, (b) pretrained on the ImageNet dataset [32], and (c) pretrained on the MS COCO dataset [31].
YOLOv3-SPP [11]
ScratchPretrained on ImagesNet Dataset [32]Pretrained on MS COCO Dataset [31]
ClassPrecisionRecallmAP0.5F1PrecisionRecallmAP0.5F1PrecisionRecallmAP0.5F1
Helmet0.92530.31440.30530.46930.93730.62750.61050.75180.82770.29710.26020.4372
Head with helmet0.92440.40350.38710.56180.93490.66680.64590.77840.78060.4630.40430.5813
person with helmet0.77780.14420.120.24330.87460.62880.59240.73160.86220.44910.40760.5906
Head0.88680.22950.21730.36460.92680.61840.59780.74180.83780.27750.24220.4169
Person without helmet0.82410.15630.13390.26280.87840.49570.47290.63380.83890.38230.35280.5252
Face0.41910.05560.02950.09820.75880.47150.42380.58160.39780.0130.0070.0252
Average0.79290.21730.19880.33330.88510.58480.55720.70320.75750.31370.2790.4294
Table 4. Comparison results of different variations of YOLOv5-P5 [13]: (a) YOLOv5s, (b) YOLOv5m, and (c) YOLOv5x.
Table 4. Comparison results of different variations of YOLOv5-P5 [13]: (a) YOLOv5s, (b) YOLOv5m, and (c) YOLOv5x.
YOLOv5s [13]YOLOv5m [13]YOLOv5x [13]
ClassPrecisionRecallmAP0.5F1PrecisionRecallmAP0.5F1PrecisionRecallmAP0.5F1
Helmet0.9610.78250.8720.86260.96320.79810.87950.87290.960.82050.88960.8848
Head with helmet0.94370.79730.87610.86080.94760.79460.87830.86410.93570.82470.89120.8767
Person with helmet0.90610.83850.89350.8710.91310.83460.89220.87210.89530.87230.90890.8836
Head0.93410.82190.8890.87440.93350.82520.88970.8760.93440.84970.90250.89
Person without helmet0.87910.75830.84930.81420.88720.76020.85270.81880.89210.79240.87320.8393
Face0.89910.65140.78630.75580.90610.69820.81220.78860.8950.74270.83010.8117
Average0.92070.7740.8610.83970.92510.78510.86870.848870.91880.8170.88260.8644
Table 5. The result of the YOLOv5x [13] and YOLOR [16] models on the publicly available datasets and the proposed SHEL5K dataset.
Table 5. The result of the YOLOv5x [13] and YOLOR [16] models on the publicly available datasets and the proposed SHEL5K dataset.
YOLOv5x [13]YOLOR [16]
DatasetsClassPrecisionRecallmAP0.5F1PrecisionRecallmAP0.5F1
SHW [28]20.93340.92970.92190.92940.94860.80630.8890.8697
Hardhat [20]30.67150.65450.63890.65460.63670.62630.64070.6315
HHW [27]30.63550.62950.62140.62880.62890.61770.63440.6233
SHD [8]30.63310.60010.5950.61280.62110.63410.64310.6276
SHEL5K60.91870.8170.88260.86440.93220.80660.88280.8637
Table 6. The result of the YOLOR [16] model on the SHEL5K dataset.
Table 6. The result of the YOLOR [16] model on the SHEL5K dataset.
YOLOR [16]
ClassPrecisionRecallmAP0.5F1
Helmet0.96580.79810.88460.874
Head with helmet0.94640.81720.88980.877
Person with helmet0.92250.87710.92040.8992
Head0.94610.84640.90680.8935
Person without helmet0.88590.80190.87670.8418
Face0.92640.69920.81820.797
Average0.93220.80660.88280.8637
Table 7. The results of K-fold cross-validation on the SHEL5K dataset using the YOLOR model [16].
Table 7. The results of K-fold cross-validation on the SHEL5K dataset using the YOLOR model [16].
K1K2K3K4K5
mAP0.5F1mAP0.5F1mAP0.5F1mAP0.5F1mAP0.5F1
Helmet0.88460.8740.88130.87040.88780.87870.8810.87020.88960.878
Head with helmet0.88980.8770.88480.87410.89320.88150.88590.87130.89530.88
person with helmet0.92040.89920.91460.89760.92130.90480.93190.91170.92260.9037
Head0.90680.89350.8930.88050.89790.8850.90680.89210.91340.9003
person without helmet0.87670.84180.87310.84330.88670.85470.87490.84120.88320.8584
face0.81820.7970.82130.79430.8140.790.80940.77950.82440.8008
Average0.88280.86370.8780.86140.88350.86580.88170.8610.88810.8714
Table 8. Results of state-of-the-art models on the SHEL5K dataset.
Table 8. Results of state-of-the-art models on the SHEL5K dataset.
ModelsPrecisionRecallmAP0.5F1Training Time
(hours)
Testing Time
(s)
Parameters
(Million)
Layers
Faster-RCNN [14]0.78080.38620.36890.516755.60.08413.348
YOLOv3-tiny [10]0.76950.42250.37790.54085.20.0068.737
YOLOv3 [9]0.85090.44820.4170.584824.60.01161.6222
YOLOv3-SPP [11]0.88510.58480.55720.703224.60.01262.6225
YOLOv4 [12]0.9250.77980.76930.844911.20.01463.9488
YOLOv4pacsp-x-mish [12]0.91950.80360.79150.856714.50.01463.9488
YOLOv5s [13]0.92050.7740.8610.83970.30.0187.1224
YOLOv5m [13]0.92510.78510.86870.84882.70.02221.1308
YOLOv5x [13]0.91880.8170.88260.86446.30.03287.2476
YOLOR [16]0.93220.80660.88280.86379.80.01236.9665
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Otgonbold, M.-E.; Gochoo, M.; Alnajjar, F.; Ali, L.; Tan, T.-H.; Hsieh, J.-W.; Chen, P.-Y. SHEL5K: An Extended Dataset and Benchmarking for Safety Helmet Detection. Sensors 2022, 22, 2315. https://doi.org/10.3390/s22062315

AMA Style

Otgonbold M-E, Gochoo M, Alnajjar F, Ali L, Tan T-H, Hsieh J-W, Chen P-Y. SHEL5K: An Extended Dataset and Benchmarking for Safety Helmet Detection. Sensors. 2022; 22(6):2315. https://doi.org/10.3390/s22062315

Chicago/Turabian Style

Otgonbold, Munkh-Erdene, Munkhjargal Gochoo, Fady Alnajjar, Luqman Ali, Tan-Hsu Tan, Jun-Wei Hsieh, and Ping-Yang Chen. 2022. "SHEL5K: An Extended Dataset and Benchmarking for Safety Helmet Detection" Sensors 22, no. 6: 2315. https://doi.org/10.3390/s22062315

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop