Next Article in Journal
Nonlinear Dynamic Analysis of a Piezoelectric Energy Harvester with Mechanical Plucking Mechanism
Previous Article in Journal
Accuracy Improvement of Braking Force via Deceleration Feedback Functions Applied to Braking Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MineSDS: A Unified Framework for Small Object Detection and Drivable Area Segmentation for Open-Pit Mining Scenario

1
Zhuzhou CRRC Times Electric Co., Ltd., Zhuzhou 412001, China
2
State Key Laboratory of Advanced Design and Manufacturing for Vehicle Body, College of Mechanical and Vehicle Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(13), 5977; https://doi.org/10.3390/s23135977
Submission received: 30 March 2023 / Revised: 3 June 2023 / Accepted: 8 June 2023 / Published: 27 June 2023
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)

Abstract

:
To tackle the challenges posed by dense small objects and fuzzy boundaries on unstructured roads in the mining scenario, we proposed an end-to-end small object detection and drivable area segmentation framework for open-pit mining. We employed a convolutional network backbone as a feature extractor for both two tasks, as multi-task learning yielded promising results in autonomous driving perception. To address small object detection, we introduced a lightweight attention module that allowed our network to focus more on the spatial and channel dimensions of small objects without impeding inference time. We also used a convolutional block attention module in the drivable area segmentation subnetwork, which assigned more weight to road boundaries to improve feature mapping capabilities. Furthermore, to improve our network perception accuracy of both tasks, we used weighted summation when designing the loss function. We validated the effectiveness of our approach by testing it on pre-collected mining data which were called Minescape. Our detection results on the Minescape dataset showed 87.8% mAP index, which was 9.3% higher than state-of-the-art algorithms. Our segmentation results surpassed the comparison algorithm by 1 percent in MIoU index. Our experimental results demonstrated that our approach achieves competitive performance.

1. Introduction

Image-based object detection is a critical task of autonomous driving perception [1], providing essential environmental information for decision-making and planning of autonomous vehicles. In mining operation situation, autonomous driving technology for mining trucks can provide valuable security guarantees for workers. However, current environment perception technology primarily aims for generalization in common scenarios and overlooks applicability for special scenarios, such as mines and ports. Designing a perception algorithm based on the particular scene’s characteristics is the foundation for improving perception in the corresponding scene. In a general mining environment, numerous small stones cluster on the surface without road lane markings, making it challenging to identify drivable areas with indistinct boundaries [2]. These issues pose significant challenges to visual perception methods, which can negatively impact mining truck driving safety. Deploying algorithms on autonomous vehicles necessitates balancing limited computational resources with low latency requirements. Compared to sequential methods, multi-task networks with a shared backbone for each task can handle tasks in parallel. Thereby, it is necessary to design an efficient object detection and drivable area segmentation joint framework for these two typical tasks of the mining scene so as to improve the perception capability of the mining truck.
Recent vision-based object detection methods, such as YOLO [3] and CornerNet [4] showed remarkable behavior in structed urban road. In relation to ordinary object detection in structed urban road, small stones are arranged more densely, irregularly, easily occluded and their appearance are similar to the background in the mining situation. Generally, small objects are defined as objects with pixels less than 32 × 32 in pictures [5]. In the past few years, small object detection (SOD) attracted the attention of researchers. Many representative studies emerged such as data-manipulation methods, feature-fusion methods, scale-aware methods and context-modeling methods [6]. In terms of data-manipulation, RRNet [7] invented AdaResampling, an adaptive resampling strategy for copying and pasting small objects in image, to improve detection accuracy. Feature pyramid network (FPN) [8] designed a bottom-up and top-down network, which can construct feature maps of different scales. FA-SSD [9] used additional features in different depth levels of the network as context, supplemented by attention mechanism to focus on the object in the image. The above methods achieved competitive results. However, there is still a large gap between the performance of these methods and that of general object detection.
In addition, drivable area segmentation for the mining environment is also a challenging task which need to be solved. The drivable area, which is entirely distinct from the structured urban road, has indefinite shape features, unstable color features and variable texture features. Recently, deep convolutional neural networks were used to handle the drivable area segmentation [10]. Some algorithms were proposed to deploy the segmentation methods directly on autonomous vehicles. For instance, ref. [11] proposed the dual-resolution networks for real-time semantic segmentation of road-driving images. Ref. [12] regarded the recognition of the drivable area as a row-selection task, which significantly reduced the computational costs. The ability of these algorithms in mining scenarios remains need to be tested. The above algorithms were designed to solve the assignments of SOD and semantic segmentation; thus, there is no general and end-to-end framework for mining environment detection and drivable area segmentation.
To summarize, we proposed a novel framework that tackled the autonomous mining vehicles challenges faced by small object detection and drivable area segmentation (also called mineSDS). Our contributions include: (1) An end-to-end framework capable of concurrently processing object detection and drivable area segmentation tasks. Our framework employs a shared feature extractor to extract features for both tasks, thereby reducing inference time and computational costs. (2) A lightweight attention module designed to enhance the detection accuracy of small objects. By enabling our network to pay more attention to small objects in the spatial and channel dimensions, we achieved improved detection performance. (3) We incorporate a convolutional block attention module to enhance the feature mapping ability of convolutional operations in the segmentation decoder, improving our network’s sensitivity to fuzzy road boundaries. Our framework’s performance was evaluated on the Minescape dataset, and our results demonstrated comparable performance in both tasks.
The remainder of this paper is organized as follows: related works are summarized in Section 2; the details of our proposed algorithm are described in Section 3. Additionally, the experimental analysis is demonstrated on Section 4. At last, the contributions of this paper are summarized on Section 5.

2. Related Works

In recent years, significant progress was made in small object detection and drivable area segmentation. This section review related work about mine scene perception, small object detection and drivable area segmentation.

2.1. Open-Pit Mining Perception

In the field of mine scene detection, Ren et al. [13] proposed a multi-scale feature fusion method for obstacle detection of roads in open pit mining areas, normalized the features, compressed the features and extracted multi-scale features. Song et al. [14] used multi-scale and attention fusion modules to capture richer context features on their own data sets and achieved good results in the field of mine target detection. For mining scene segmentation, Wei et al. [15] extracted multi-scale features by a multi-branch feature extraction network. Xiao et al. [16] designed RATT-UNet network to extract open-pit mine roads information. Liu et al. [17] used point cloud data to track mine road.

2.2. Small Object Detection

Small object detection made great progress in recent years. The researchers mainly made improvements to solve problems as follows: (1) the feature information of small object is extremely lacking, (2) small gravel objects are more densely arranged, more irregular, and more easily blocked. First of all, many studies made improvements in the data-manipulation methods. The most direct way is to increase the number of small objects in the data set. Ref. [18] first proposed the strategy of randomly copying small objects in the data set, simply and directly increasing the number of small objects. On this foundation, Chen et al. proposed the [7] network to determine the background of small objects before copying and pasting small object samples. In addition, Duan, Wei et al. [19,20] proposed a variety of ways to use mosaic to solve the small object detection problem, and achieved certain results. The mosaic data augmentation method was proposed in YOLOv4 [21] paper. The main idea is to randomly cut four pictures and then splice them into one picture as training data. The advantage of this is that it enriches the background of image, and the four images are spliced together to improve the batch size in a disguised way, four images will also be calculated during batch normalization.
For small objects, the most important thing is to complete the feature extraction at the corresponding scale. Dollár [8] used an image pyramid with a sliding window scheme to extract features. Yang et al. [22] selected a feature layer of appropriate size for the pooling operation of small object features. Cai et al. proposed the MS-CNN [23] method to select the best reception region for small object objects in different layer scales. In the solution of feature fusion, Lin et al. [8] first proposed a feature pyramid network (FPN), the purpose of which is to aggregate high resolution semantic features with higher level features in low resolution. This simple and effective design became an important part of the feature extractor. Zand et al. built DarkNet RI [24] on the basis of DarkNet-53 and skip-connection to generate high-level semantic feature maps of different scales.
Using attention mechanism is also a common way for small object detection. Its main principle is to use the weight relationship to obtain the feature information that should pay more attention. SE Attention [25] module is the comparative attention, which is characterized by adding attention mechanism in the dimension of special channel. ECA-Net [26] module effectively captures cross channel interaction by using one-dimensional convolution. CBAM (convolutional block attention module) [27] starts from the two scopes of channel and spatial, introduces two analysis dimensions of spatial attention and channel attention and realizes the sequential attention structure from channel to space.

2.3. Drivable Area Segmentation

Semantic segmentation involves pixel-level classification of dense image prediction. The problem of early semantic segmentation is that it cannot effectively extract high-level semantic features, and the effect is poor on complex images. The fixed receptive field cannot match the scale change of the object, and simple up-sampling will lead to the loss of detailed information. Long et al. first proposed a network (FCN) [28] that can process features on images of arbitrary resolution. U-Net [29] proposed a novel u-shape network structure, in which the down-sampling image features are fused with the up-sampling information and the pixel level features are segmented to avoid the loss of pixel level semantic features. Chen et al. proposed DeepLab [30,31], which first used dilation convolution to increase receptive field, effectively avoiding the problem of feature disappearance in down-sampling.
Drivable area segmentation derives from semantic segmentation. Drivable area segmentation task is a special task of semantic segmentation. Travelable area segmentation is mainly used for the perception of autonomous vehicle. Due to the particularity of the task and the characteristics of large range and clear edge features of the drivable area compared with complex segmentation targets, drivable area segmentation requires more accuracy and real-time. In the aspect of drivable area segmentation, the YOLOP [32] panoramic perception network completed three times of up-sampling on the basis of YOLO framework to achieve the segmentation of drivable area. YOLOPv2 [33] uses a new mixed loss function to improve the segmentation and detection effects. HybridNets [34] uses the efficient segmentation head of the weighted bidirectional feature network to perform segmentation, which is an accurate and practical method. Hong et al. proposed DDRNet [11], which uses deep dual-resolution networks are composed of two deep branches between which multiple bilateral fuels are performed. Asgarian et al. [12] consider the process of drivable area recognition as a row-selection task, increasing the speed of performance. Wang et al. added C6 and C7 modules with larger receiving fields on the basis of YOLACT [35] to improve the segmentation of the drivable area.
In order to better detect small obstacles on the road surface, we presented two networks: (1) we added a combination of SimAM and channel attention CAM (channel attention module) module on the last layer of the in the redesigned YOLOv7 head, which achieved better detection results. (2) In terms of segmentation, five up-samples followed by RepbottleneckCSPA and CBAM were used. As for the above two points, we proposed MineSDS, which can simultaneously meet the requirements of obstacle detection in the mining area and segmentation of the exercisable area.

3. Methodology

We put forward a framework to achieve small object detection and semantic segmentation; the details of the proposed method are shown in Figure 1. As the figure shows, our architecture consisted of a (a) feature extractor, (b) segment decoder and (c) multi-head feature fused detector.

3.1. Feature Extractor

The feature extraction module takes an image as input and outputs semantic features, which can be used as input for subsequent object detection and semantic segmentation. Many representative classification networks were used for feature extractors, such as ResNet [36], DenseNet [37], etc. In the down-sample module, we used both max pooling and convolution with the stride of two. The input image first passed through three convolution modules, and the size of feature map became ( H / 4 ,   W / 4 ,   128 ) . Then, the feature went into three modules consisting of ELAN module and down-sample module. Finally, the feature was enhanced by an ELAN module. The input image, thus, became a feature map with the size of ( H / 32 ,   W / 32 ,   1024 ) . The detail of the feature extractor is shown in Figure 2:

3.2. Multi-Head Feature Fused Detector

We implemented a multiscale detection scheme in our multi-head detection module, which relied on anchor-based detection. To aggregate features, we used a combination of feature pyramid network (FPN) and path aggregation network (PAN) modules. The feature map was first passed through a top-down feature pyramid, and then through a bottom-up feature pyramid with PAN structure. This allowed us to merge information from different scales and perform object detection at different scales using multiple prediction heads directly on the fused feature map of the PAN. In our approach, each grid on the feature map corresponded to a branching head, which was assigned three anchors of different shapes. Each detection head predicted three tensors encoding the bounding box, category and prediction confidence. By using FPN and PAN, we can fuse features at different scales, which allows the smaller object detection branch to obtain valuable information from earlier feature maps, preventing the loss of information during repetitive down sampling. Furthermore, we proposed a small object attention submodule to enhance detection of smaller objects. This submodule reassigned feature weights to better detect small objects on the branch that detects smaller objects.
In order to address the challenges associated with detecting small objects, we introduced a novel small scale-wise attention module that was capable of reassigning weights for both spatial and channel features. This innovative module comprised two components: SimAM and CAM.
SimAM is a method grounded in neuroscience theory, which employs energy functions to explore the significance of various elements within the feature map. Utilizing SimAM, we were able to derive 3D attention, whereby spatial attention was 2D and channel attention was 1D. The energy function was initially defined in thermodynamics to characterize the stability of a system. It is now widely employed to express the degree of divergence of elements within a system, where smaller energy values correspond to a more stable system. The energy function can be mathematically expressed as:
e v w v , b v , u , x i = u v v ^ 2 + 1 L 1 i = 1 L 1 u o x ^ i 2
where v ^ = w v v + b v represents the feature values of the target elements and x i ^ = w v x i + b v , which represents the feature values of other elements in the same spatial dimension. i and L = W × H is the index of the feature element and the number of feature element over spatial dimension.
To find the linear separability between the target element v ^ and other element x ^ i , it adopts binary labels (i.e., 1 and −1) for u v and u o . and also add a regularization term into Equation (1) for simplicity. Then, the final energy function is as follows:
e v w v , b v , u , x i = 1 L 1 i = 1 L 1 1 w v x i + b v 2 + 1 w v v + b v 2 + λ w v 2
By minimizing the final energy function, the separability between v ^ and x ^ i can be found. Additionally, the minimal energy can be computed with the following:
e v min = 4 σ ^ 2 + λ ( v μ ^ ) 2 + 2 σ ^ 2 + 2 λ
here, μ ^ = 1 l i = 1 L 1 x i and σ ^ 2 = 1 L i = 1 L 1 x i μ ^ 2 the smaller e v min , the more different the element is from the surrounding elements and the higher the corresponding attention weight. We can achieve attentional weight redistribution by:
X = sigmoid 1 E X
where X , X R C × H × W .
With this method, the spatial information of the small object was well attended to by the network.
It was observed that while the attention weights generated by SimAM were 3D, the majority of the attention was assigned to the spatial dimension. To address the reassignment of channel weights to features, which is a common practice in attention methods, we implemented the CAM prior to SimAM. The CAM comprised two parallel branches that performed max-pooling and avg-pooling, respectively. The channel attention weight was then obtained through MLP and normalized using the sigmoid function, before being used to weight each channel in the original input feature map one by one to recalibrate the channel attention to the original feature. The overall channel attention process can be summarized as follows:
A c F in = σ ( MLP MaxPool F in + MLP AvgPool F in )
F out = A c F i n F i n
F in and F out are the input feature map and output feature map, F in , F out R C × H × W . MLP is multilayer perceptron.

3.3. Segmentation Decoder

The segmentation decoder is responsible for assigning a label (drivable area or background) to each pixel in the input image. The input feature map with a size of “(H/32, W/32, 1024)” was fed into the segmentation decoding module. After performing five up-sampling module operations, the output feature map was restored to the original image size, representing the probabilities of each pixel being part of the background or drivable area. Our up-sampling module included an up-sampling operation and a RepBottleneckCSP module, which employed the CSPNet architecture to connect rep convolution groups. The CSP architecture enabled richer gradient combinations while reducing computational complexity, and the RepConv structure improved fitting ability by using a multichannel structure during training and a single-channel structure during inference, making the RepBotterneckCSP module a popular choice for many mainstream networks. The structure of RepBotterneckCSP is shown in Figure 3.
To address the challenging problem of the fuzzy boundary of drivable areas, we used the CBAM module to redistribute feature weights after the first up-sampling operation as shown in Figure 4. This was achieved by inferring the attention map along two independent dimensions (channel and space) and then multiplying it with the input feature map to perform adaptive feature optimization. This step was necessary to enable the segmentation decoding module to focus more on the fuzzy boundary of drivable areas.

3.4. Loss Function

Our framework performed object detection and drivable area segmentation simultaneously, and our loss function included the object detection loss and the drivable area segmentation loss.
The object detection loss L det contains classification loss L cls , object loss L obj and bounding box loss L bbx . L cls and L obj are both binary cross-entropy loss, which is used to minimize the error of classification and of the confidence of one prediction separately. We used SIoU loss [38] as L bbx , which takes the distance, overlap rate, the similarity of shape and orientation between the predicted box and ground truth into account, while most other loss functions do not consider orientation.
The definition of SIoU loss is:
Loss SIoU = 1 IoU + Δ + Ω 2
The angle cost Λ is:
Λ = 1 2 sin 2 arcsin ( x ) π 4
The distance cost Δ is:
Δ = t = x , y 1 e γ ρ t
where γ = 2 Λ .
The shape cost Ω is
Ω = t = w , h 1 e ω t θ
The IoU cost is:
IoU = B B GT B B GT
As Equation (12), the L det is the weight sum of three loss above.
L det = α 1 L cls + α 2 L obj + α 3 L bbx
The drivable area segmentation loss L seg is binary cross-entropy loss, which is used to penalize the classification of each pixel in th Modified e segmentation map.
In conclusion, the final loss function of our network is given by.
L = α L det + β L seg
where α and β are the weight of detection loss and segmentation loss.
We can tune α 1 , α 2 , α 3 to balance the three parts of detection task and tune α , β to balance the detection part and segmentation part.

4. Experiment

In this section, we demonstrate the details of our experiment and that all our experiments in this paper were conducted based on Intel Xeon E5-2698 v4 and one 32G NVIDIA TESLA V100. The construction of the experimental model was based on PyTorch.

4.1. Dataset

Currently, there are limited datasets for object detection and segmentation specifically tailored for mining scenes. To validate our algorithm, we constructed our own mining perception dataset called Minescape. According to the actual mining operation scenario, we used a XDE240 mining truck as the data collection platform and its FLIR Point Grey Flea front camera as the image sensor with 10Hz capture frequency as shown in Figure 5. The Minescape consisted of 4547 training images, 1299 verification images and 650 test images with various mine scenes, 60+ driving hours of three days and varying luminance conditions from Shenyan Coal and Xiwan Opencast Coal Mines. To ensure the accuracy of annotations, we used labelme to manually annotate the images. We evaluated our network on the Minescape verification set.
The Minescape had 9 types of label objects, including 4 types of movable objects (mining truck, car, person, engineering vehicle) and 5 types of static objects (stone, other obstacle, sign, background, drivable area), as shown in Table 1. These categories cover most of the work scenes of mining trucks, and their scale and number distribution is very uneven, which is one of the reasons why the general detection framework is not robust in the mining environment.
It is worth mentioning that the advantage of the Minescape is that it reveals some common problems in open-pit mine, such as clustered small objects and the unobvious boundary of drivable areas, as shown in Figure 6. The Minescape provided us with data in various typical scenarios to design a framework to solve such problems.

4.2. Metrics

The algorithm proposed by us included two tasks: detection and segmentation. A general evaluation index can judge the performance of the model, which is also a visual manifestation of the comparison and evaluation with existing models. At present, the evaluation indexes of target detection algorithm widely accepted by many scholars include Loss Function, Intersection over Union (IoU) and mean Average Accuracy (mAP). The evaluation indicators of segmentation tasks mainly include Mean Intersection over Union (MIoU) and Accuracy (Acc).

4.2.1. Evaluation Metrics of Target Detection

The intersection over union (IoU) is an important reference index in the field of target detection, which represents the overlap rate between candidate boxes and tag boxes. When IoU value is 1, it indicates that candidate box and marker box are completely coincident, and the formula is as follows:
IoU = A B A B
Precision refers to the proportion of real positive cases among the samples predicted as positive cases, expressed as P. Formula is shown as follows:
p = TP TP + FP
Recall rate refers to the proportion of correctly predicted samples in all actually positive samples. The formula is as follows:
recal = TP TP + FN
Average Precision (AP) means the average precision rate under different recall rates, expressed as the formula:
AP = i = 1 n P ( i ) r ( i ) = 0 1 p ( r ) dr
The definition of Mean Average Precision (mAP) is to calculate the mean of AP, which is the average precision of all categories. The formula is as follows:
mAP = n = 1 N AP ( n ) N
in the formula, represents the category, and represents the total number of categories.

4.2.2. Evaluation Metrics for Task Segmentation

IoU represents the ratio of the intersection and union of the two sets of the real area and the prediction area, and MIoU represents to calculate the IoU of each category, then sum and average. The calculation formula of MIoU is:
MIoU = 1 k i = 1 κ V Ti V Pi V Ti V Pi
Pixel Accuracy represents the percentage of correctly classified pixels in the image. It’s given by:
Accuracy = TP + TN TP + TN + FT + FN

4.3. Implementation Detail

We used SGD with momentum as the optimizer, with an initial learning rate of 0.01. To better train our model, we performed 3 epochs of warm up with an initial learning rate of 0.1. We trained 80 epochs on our dataset with the batch size of 16 and the input image size of ( 640 ,   640 ,   3 ) . In the final loss function settings, the parameters α 1 , α 2 , α 3 , α and β as mentioned before were 0.5, 2.0, 0.1, 1.0 and 0.8, respectively.

4.4. Experiment Results

We developed an end-to-end multi-task network. However, it was difficult to find a single model that can fulfill all multi-task requirements. Currently, a model that performs well in object detection may not perform well in segmentation. Therefore, we compared the performance of our model with the state-of-the-art (SOTA) algorithms both in object detection and semantic segmentation. The results are shown in Table 2.
To evaluate small object detection, we compared our model with SOTA object detection algorithms such as YOLOv7tiny, YOLOv5n and Faster-RCNN. Our dataset consisted of small objects such as stones, signs and other obstacles, and we reported mAP results for the above algorithms, which ranged from 32.2% to 73.3%. Our model achieved significantly higher results of 67.6%/87.5%/86.5%, with the highest improvement of 27.0%/18.5%/13.2% over the SOTA algorithms. Interestingly, our model used the same backbone as YOLOv7tiny, but with the addition of Multi-head Feature Fused Detector and CAM-SimAM, we were able to extract better small object features.
We also conducted a comparison experiment of segmentation networks using the same experimental setup. We compared three networks: Bisnet, U-Net and HarDNet, which are the baseline algorithms for classic segmentation networks and drivable area segmentation algorithms. We used two evaluation indicators, MIoU and Acc, and our algorithm achieved higher results of 0.6850/0.8150, which were 1% higher than each other. The results are shown in Table 3.
We visualized the prediction results of our model on Minescape, and some of the results are shown in Figure 7. It can be seen that our model can detect the objects very well, and it also showed comparatively well for the small objects, but there were still, sometimes, missing detections when comparing with the ground truth. In terms of segmentation, comparing with the ground truth, it can basically segment the boundary correctly, and the segmentation effect became worse at the farther distance.

4.5. Ablation Study

The purpose of conducting ablation experiments is to explore the contribution of various components within our framework. We conducted corresponding experiments to explore the role of attention components and loss functions. Additionally, we trained our model to perform both object detection and drivable area segmentation tasks separately to assess the effectiveness of our multi-task training approach.
Our experiments and their corresponding results are presented in Table 4. The results demonstrate that the attention module significantly improved the model’s overall mAP by 2.9%, with significant improvements seen in small objects such as stones, other obstacles and signs. On the other hand, there was no significant difference observed in the improvement of other categories with normal scale. Our multi-task training framework resulted in a 1.5% increase in overall mAP compared to training solely for the detection task. However, the difference between the result of training for the segmentation task individually and that of multi-task training was not apparent. Experimentally, we found that the SIoU loss function was more favorable to our model’s training than CIoU, with the model trained with SIoU yielding a 2.2% higher mAP than the model trained with CIoU.

5. Conclusions

In this paper, we proposed a perceptual framework for mining scenarios that focused on improving small object detection and drivable area segmentation. To achieve this, we employed a small object attention module that enhanced the network’s accuracy for detecting numerous small objects. Additionally, we used CBAM to improve the feature mapping capabilities of convolutional operations, allowing our network to focus on the fuzzy boundaries of drivable areas. We also trained both tasks simultaneously using a weighted loss function. Our extensive experiments demonstrated that our framework performed comparably well. However, our work had some limitations. Firstly, the data we collected did not include bad weather conditions, which could impact the robustness of the algorithm in real-world applications. To address this, we will need to re-collect a substantial amount of data. Moreover, we did not operate our algorithm and other algorithms on a classical benchmark dataset. This may be biased. We will compare our algorithm with state-of-the-art methods on a more representative benchmark dataset.

Author Contributions

Conceptualization, J.H. and M.G.; methodology, M.G. and C.L.; software, C.L. and Y.L.; validation, C.L.; data curation, Y.L.; writing—original draft preparation, J.H.; writing—review and editing, J.H. and M.G.; visualization, C.L. and Y.L.; supervision, M.G.; funding acquisition, M.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research funded by the National Key R&D Program of China (2021YFB2501800), National Natural Science Foundation of China (52102461), and State Key Laboratory of Advanced Design and Manufacturing for Vehicle Body (32115013).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to third party (data source) restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Balasubramaniam, A.; Pasricha, S. Object Detection in Autonomous Vehicles: Status and Open Challenges. arXiv 2023, arXiv:2201.07706. [Google Scholar]
  2. Li, Y.; Li, Z.; Teng, S.; Zhang, Y.; Zhou, Y.; Zhu, Y.; Cao, D.; Tian, B.; Ai, Y.; Xuanyuan, Z.; et al. AutoMine: An Unmanned Mine Dataset. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–20 June 2022; pp. 21276–21285. [Google Scholar]
  3. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2023, arXiv:2207.02696. [Google Scholar]
  4. Law, H.; Deng, J. CornerNet: Detecting Objects as Paired Keypoints. In Proceedings of the Computer Vision—ECCV, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 765–781. [Google Scholar]
  5. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  6. Cheng, G.; Yuan, X.; Yao, X.; Yan, K.; Zeng, Q.; Xie, X.; Han, J. Towards Large-Scale Small Object Detection: Survey and Benchmarks. arXiv 2023, arXiv:2207.14096. [Google Scholar]
  7. Chen, C.; Zhang, Y.; Lv, Q.; Wei, S.; Wang, X.; Sun, X.; Dong, J. RRNet: A Hybrid Detector for Object Detection in Drone-Captured Images. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 100–108. [Google Scholar]
  8. Dollár, P.; Appel, R.; Belongie, S.; Perona, P. Fast Feature Pyramids for Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1532–1545. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Jiang, D.; Sun, B.; Su, S.; Zuo, Z.; Wu, P.; Tan, X. FASSD: A Feature Fusion and Spatial Attention-Based Single Shot Detector for Small Object Detection. Electronics 2020, 9, 1536. [Google Scholar] [CrossRef]
  10. Oliveira, G.L.; Burgard, W.; Brox, T. Efficient deep models for monocular road segmentation. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 4885–4891. [Google Scholar]
  11. Pan, H.; Hong, Y.; Sun, W.; Jia, Y. Deep Dual-Resolution Networks for Real-Time and Accurate Semantic Segmentation of Traffic Scenes. IEEE Trans. Intell. Transp. Syst. 2023, 24, 3448–3460. [Google Scholar] [CrossRef]
  12. Asgarian, H.; Amirkhani, A.; Shokouhi, S.B. Fast Drivable Area Detection for Autonomous Driving with Deep Learning. In Proceedings of the 2021 5th International Conference on Pattern Recognition and Image Analysis (IPRIA), Kashan, Iran, 28–29 April 2021; pp. 1–6. [Google Scholar]
  13. Ren, L.; Yang, C.; Song, R.; Chen, S.; Ai, Y. An Feature Fusion Object Detector for Autonomous Driving in Mining Area. In Proceedings of the 2021 International Conference on Cyber-Physical Social Intelligence (ICCSI), Beijing, China, 18–20 December 2021; pp. 1–5. [Google Scholar]
  14. Song, R.; Ai, Y.; Tian, B.; Chen, L.; Zhu, F.; Fei, Y. MSFANet: A Light Weight Object Detector Based on Context Aggregation and Attention Mechanism for Autonomous Mining Truck. IEEE Trans. Intell. Veh. 2022, 8, 2285–2295. [Google Scholar] [CrossRef]
  15. Wei, Q.; Song, R.; Yang, X.; Ai, Y. A real-time semantic segmentation method for autonomous driving in surface mine. In Proceedings of the 2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC), Macau, China, 8–12 October 2022; pp. 2763–2768. [Google Scholar]
  16. Xiao, D.; Yin, L.; Fu, Y. Open-Pit Mine Road Extraction From High-Resolution Remote Sensing Images Using RATT-UNet. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  17. Lu, X.; Ai, Y.; Tian, B. Real-Time Mine Road Boundary Detection and Tracking for Autonomous Truck. Sensors 2020, 20, 1121. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar]
  19. Wei, Z.; Duan, C.; Song, X.; Tian, Y.; Wang, H. AMRNet: Chips Augmentation in Aerial Images Object Detection. arXiv 2020, arXiv:2009.07168. [Google Scholar]
  20. Duan, C.; Wei, Z.; Zhang, C.; Qu, S.; Wang, H. Coarse-grained Density Map Guided Object Detection in Aerial Images. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 2789–2798. [Google Scholar]
  21. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  22. Yang, F.; Choi, W.; Lin, Y. Exploit All the Layers: Fast and Accurate CNN Object Detector with Scale Dependent Pooling and Cascaded Rejection Classifiers. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, Nevada, 26 June–1 July 2016; pp. 2129–2137. [Google Scholar]
  23. Cai, Z.; Fan, Q.; Feris, R.S.; Vasconcelos, N. A Unified Multi-scale Deep Convolutional Neural Network for Fast Object Detection. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 27–30 June 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 354–370. [Google Scholar]
  24. Zand, M.; Etemad, A.; Greenspan, M. Oriented Bounding Boxes for Small and Freely Rotated Objects. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4701715. [Google Scholar] [CrossRef]
  25. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 8 June 2018; pp. 7132–7141. [Google Scholar]
  26. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar]
  27. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision—ECCV, Munich, Germany, 8 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar]
  28. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–15 June 2015; pp. 3431–3440. [Google Scholar]
  29. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  30. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  32. Wu, D.; Liao, M.-W.; Zhang, W.-T.; Wang, X.-G.; Bai, X.; Cheng, W.-Q.; Liu, W.-Y. YOLOP: You Only Look Once for Panoptic Driving Perception. Mach. Intell. Res. 2022, 19, 550–562. [Google Scholar] [CrossRef]
  33. Han, C.; Zhao, Q.; Zhang, S.; Chen, Y.; Zhang, Z.; Yuan, J. YOLOPv2: Better, Faster, Stronger for Panoptic Driving Perception. arXiv 2022, arXiv:2208.11434. [Google Scholar]
  34. Vu, D.; Ngo, B.; Phan, H. HybridNets: End-to-End Perception Network. arXiv 2022, arXiv:2203.09035. [Google Scholar]
  35. Bolya, D.; Zhou, C.; Xiao, F.; Lee, Y.J. YOLACT: Real-Time Instance Segmentation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; IEEE: Seoul, Republic of Korea, 2019; pp. 9156–9165. [Google Scholar]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Las Vegas, NV, USA, 2016; pp. 770–778. [Google Scholar]
  37. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: Honolulu, HI, USA, 2017; pp. 2261–2269. [Google Scholar]
  38. Gevorgyan, Z. SIoU Loss: More Powerful Learning for Bounding Box Regression. arXiv 2022, arXiv:2205.12740. [Google Scholar]
Figure 1. The architecture of MineSDS. (a) Feature Extractor. (b) Multi-Head Feature Fused Detector. (c) Segmentation Decoder.
Figure 1. The architecture of MineSDS. (a) Feature Extractor. (b) Multi-Head Feature Fused Detector. (c) Segmentation Decoder.
Sensors 23 05977 g001
Figure 2. Feature Extractor.
Figure 2. Feature Extractor.
Sensors 23 05977 g002
Figure 3. RepBotterneckCSP.
Figure 3. RepBotterneckCSP.
Sensors 23 05977 g003
Figure 4. Convolutional Block Attention module.
Figure 4. Convolutional Block Attention module.
Sensors 23 05977 g004
Figure 5. Data collection platform and the position of front camera.
Figure 5. Data collection platform and the position of front camera.
Sensors 23 05977 g005
Figure 6. Some representative pictures in Minescape. (a) The distribution of small objects in the mine. They are either aggregated or dispersed. They are arranged irregularly and have strong camouflage. (b) The characteristics of the drivable area. The boundary shape of the drivable area is irregular and sometimes discontinuous. Its color and texture features are also unstable.
Figure 6. Some representative pictures in Minescape. (a) The distribution of small objects in the mine. They are either aggregated or dispersed. They are arranged irregularly and have strong camouflage. (b) The characteristics of the drivable area. The boundary shape of the drivable area is irregular and sometimes discontinuous. Its color and texture features are also unstable.
Sensors 23 05977 g006
Figure 7. Results on Minescape.
Figure 7. Results on Minescape.
Sensors 23 05977 g007
Table 1. Detail of classes in Minescape.
Table 1. Detail of classes in Minescape.
GroupClassQuantityDescription
movable objectsmining truck4845mining dump trucks
car601cars and pickups
person2965workers
engineering vehicle5638excavator, etc.
static objectsstone136,578ore and rock
sign1205traffic signs
other obstacles2575other common obstacles
drivable area 9376safe drivable ground
background3035undrivable area
Table 2. Evaluation results of different detection algorithms on Minescape.
Table 2. Evaluation results of different detection algorithms on Minescape.
ModelBackbonemAP0.5 (%)
AllMining
Truck
Other
Obstacle
CarPersonEngineering
Vehicle
StoneSign
YOLOv7
tiny
ELAN78.590.870.489.495.696.940.666.0
YOLOv5nCSPNet77.388.469.589.189.495.340.469.0
Faster-RCNNResNet-5076.887.873.392.589.195.532.267.0
MineSDS-det (ours)ELAN87.895.786.594.994.998.367.687.5
Table 3. Evaluation results of different segmentation algorithms on Minescape.
Table 3. Evaluation results of different segmentation algorithms on Minescape.
NetworkMloUAccuracy
Bisnet0.64500.7842
U-Net0.65710.7931
HarDNet0.67750.8079
MineSDS-seg (ours)0.68500.8150
Table 4. Ablation studies of MineSDS.
Table 4. Ablation studies of MineSDS.
MethodmAP0.5 (%)mloU (%)Accuracy
AllMining
Truck
Other
Obstacle
CarPersonEngineering
Vehicle
StoneSign
X (base)84.995.780.995.194.397.752.677.668.281.3
Y (X + SSWAM)87.895.786.594.993.998.357.687.568.581.5
Y-Det (only)86.395.283.293.993.897.855.884.4--
Y-Seg (only)68.481.5
CIoU Loss77.690.473.487.693.196.344.857.768.381.4
SIoU Loss81.792.679.590.794.397.545.871.768.481.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Li, C.; Huang, J.; Gao, M. MineSDS: A Unified Framework for Small Object Detection and Drivable Area Segmentation for Open-Pit Mining Scenario. Sensors 2023, 23, 5977. https://doi.org/10.3390/s23135977

AMA Style

Liu Y, Li C, Huang J, Gao M. MineSDS: A Unified Framework for Small Object Detection and Drivable Area Segmentation for Open-Pit Mining Scenario. Sensors. 2023; 23(13):5977. https://doi.org/10.3390/s23135977

Chicago/Turabian Style

Liu, Yong, Cheng Li, Jiade Huang, and Ming Gao. 2023. "MineSDS: A Unified Framework for Small Object Detection and Drivable Area Segmentation for Open-Pit Mining Scenario" Sensors 23, no. 13: 5977. https://doi.org/10.3390/s23135977

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop