Next Article in Journal
A Blockchain-Based Trustworthy Model Evaluation Framework for Deep Learning and Its Application in Moving Object Segmentation
Previous Article in Journal
Automated Gait Analysis Based on a Marker-Free Pose Estimation Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MFF-YOLO: An Accurate Model for Detecting Tunnel Defects Based on Multi-Scale Feature Fusion

School of Electronic Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450045, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6490; https://doi.org/10.3390/s23146490
Submission received: 9 June 2023 / Revised: 12 July 2023 / Accepted: 13 July 2023 / Published: 18 July 2023
(This article belongs to the Topic Artificial Intelligence in Sensors, 2nd Volume)

Abstract

:
Tunnel linings require routine inspection as they have a big impact on a tunnel’s safety and longevity. In this study, the convolutional neural network was utilized to develop the MFF-YOLO model. To improve feature learning efficiency, a multi-scale feature fusion network was constructed within the neck network. Additionally, a reweighted screening method was devised at the prediction stage to address the problem of duplicate detection frames. Moreover, the loss function was adjusted to maximize the effectiveness of model training and improve its overall performance. The results show that the model has a recall and accuracy that are 7.1% and 6.0% greater than those of the YOLOv5 model, reaching 89.5% and 89.4%, respectively, as well as the ability to reliably identify targets that the previous model error detection and miss detection. The MFF-YOLO model improves tunnel lining detection performance generally.

1. Introduction

With China’s recent economic growth, the tunnel sector has also entered a golden age of development, and tunnel building has emerged as one of the crucial elements of China’s infrastructure development. However, due to geological factors, poor construction practices, and natural calamities, tunnels may have concealed flaws including uncompacted, hollow, and water filling that gravely jeopardize their service life [1].
To identify and address issues with the tunnels as soon as they arise, regular inspection and maintenance are required. However, conventional methods for finding tunnel defects, such as visual inspection and acoustic inspection, have low detection efficiencies and high result error. Given the quick advancement of deep learning technologies in the detection of targets, such as damage detection and localization of bridge deck pavement [2], crack detection in concrete bridges [3,4], steel surface flaw detection [5,6], and wheel defect detection [7] across a variety of disciplines in society, science, and engineering, deep learning-based tunnel flaw detection has recently gained the attention of both domestic and international academics. For example, Sjölander et al. [8] summarized the research on the application of optical detection technology and autonomous evaluation methods based on machine learning technology in tunnel lining inspection, as well as the research on digital cameras, laser scanning, fiber optic sensors, and other methods; they also proposed issues with traditional tunnel inspection methods, such as their low efficiency. Deep learning technology was used by Maeda et al. [9] to address the issue of data acquisition and traditional data enhancement techniques like rotation, translation, and flip that may alter the semantic information of the image and cause other drawbacks, using methods like selective image cropping and stitching to address the issue of the insufficient data set; however, the model detection effect did not improve. With the primary objective of resolving the data signal-to-noise ratio and multi-path interference problems to offer assurances for data feature extraction, Lei et al. [10] suggested an air-coupled geo-radar detection technique provided by F-K filtering and BP migration. A strong technical guarantee for single defect detection was provided by the GPR forward simulation model developed by Wu et al. [11], but since tunnel defects are frequently made up of hollow, water-filling, and non-compactness, the model’s generalization ability is not ideal. Ali et al. [12], using Faster-RCNN and YOLOv3 networks as well as conventional detection techniques, examined the performance of concrete structures; the findings revealed that the convolutional neural network-based detection strategy had greater detection accuracy and localization precision. Before training the model with an enhanced U-Net network, Wang et al. [13] first performed picture preprocessing. Smaller and more subtle faults might be detected by the model, but the model’s efficiency was low, and its real-time performance could not keep up with demand. Li et al. [14] used U-Net and alternating update convolutional neural networks for automatic tunnel defect detection: first, the image is segmented and predicted to extract defect features; next, alternating update convolutional neural networks are used for classification and localization; however, the model is more complex and cannot satisfy the tunnel defect detection engineering requirements.
Although several researchers have developed numerous techniques for finding tunnel defects, these techniques still have flaws and disadvantages, such as low detection accuracy and slow detection speeds. Therefore, we are working on the relevant research and attempting to address the current shortcomings to further enhance the effectiveness and accuracy of tunnel defect detection. To increase the accuracy and usefulness of tunnel defect detection, as well as to better support and guarantee tunnel construction and maintenance operations, we will keep researching new methods and approaches. Previous studies by our group have included improvements to SGD networks and residual modules [15], the introduction of adversarial networks for data expansion problems [16], and the use of neural network fusion techniques to improve the generalization performance of the model.
This study addresses the current issues with tunnel defect detection in two ways: first, in terms of data acquisition, the most recent radar detection equipment’s working process is shown in Figure 1, which has the qualities of high accuracy and high resolution and can precisely identify the defects within the depth of 10 m underground through the on-site survey; second, in terms of data processing, a detection model based on multi-scale feature fusion technology is proposed in this study, which uses the MFF-YOLO method to detect defects.
The following are this paper’s key contributions:
(1) Designing the Weighted Cross Connections Feature Pyramid Network (WCFPN) to address the issue of missing feature map information.
(2) Designing the Re-weighted Non-Maximum Suppression (RWNMS) to address the issue of redundancy and miss detection of detection frames.
(3) Improving the loss function by integrating the aspect ratio and Euclidean distance factors to improve the model training efficiency and detection accuracy.
The remainder of the text is organized as follows: Section 2 discusses the work on multiscale feature fusion and convolutional neural networks; Section 3 describes the technological strategies for improvement; the data set and assessment metrics are introduced in Section 4; tests are carried out in Section 5 to confirm the improved model performance; and the work is reviewed and expected in Section 6.

2. Related Work

2.1. Convolutional Neural Network

CNN [17] is a deep learning model consisting of convolutional, pooling, and fully connected layers, etc., for implementing tasks such as image classification, target detection, and image segmentation. One of the target detection methods can be divided into two-stage and single-stage, and the difference between the two algorithms lies in the way the generation frame is combined with the candidate frame. In a two-stage algorithm, candidate regions are first generated and then the CNN applies to these candidate regions for classification. For example, the R-CNN [18,19,20,21,22] family of algorithms has high detection accuracy but is computationally intensive and inefficient. In contrast, single-stage algorithms such as the SSD [23,24,25,26,27] and YOLO [28,29,30,31,32,33,34,35] series can achieve significant detection speedups. Single-stage target detection algorithms have become preferred in industrial applications due to their ability to directly output information about the position and detection frame of the target to be detected. In conclusion, with the continuous development of deep learning techniques, target detection algorithms are also constantly being advanced and optimized, with different algorithms being suitable for different application scenarios and needs.

2.2. Multi-Scale Feature Fusion

Multi-scale feature fusion is a technique for combining feature maps at different scales to improve the performance of computer vision tasks. Common approaches include cascade structures [36,37,38], pyramid networks [39,40,41,42] and attention mechanisms [43,44,45]. Cascade structures link feature maps at different scales together to form cascade networks, which can be effective in improving performance; pyramid networks are a hierarchical approach to image processing that extracts features at different scales and combines them; and attention mechanisms can make the network focus more on important features by weighting feature maps at different scales. All these methods aim to improve the performance of computer vision tasks by using feature maps at different scales, thus improving the accuracy and robustness of the task.

3. Methods

The MFF-YOLO model improvement approach is described in this section, and Figure 2 illustrates the structural characteristics of the MFF-YOLO model. The model improvement consists of two main components: the WCFPN structure of the neck network, where the blue module represents the Weighted Cross Connections built by weighting and the red line portion reflects the network across connection idea; and the RWNMS mechanism at the predicted end, where the EIOU loss function is used in the screening process of the prediction frame and the weights are further provided by normalization.

3.1. WCFPN

The target detection model is prone to the problem that as the number of layers in the feature map increases, the resolution of the feature map decreases, resulting in poor information transfer in the feature map and hence the lack of detection capability of the model. The model may zoom in and out of the feature maps by adding operations like convolutional layers and pooling layers; however, this reduces the performance of the model since they commonly cause information loss or information redundancy. The study uses cross-scale linking to connect feature maps of multiple sizes, better using the data in those feature maps, and then suggests the concept of weighted fusion when zooming in and out on the feature maps to address this issue.
Using the effective feature fusion technique known as weighted fusion, data are combined in feature maps of various scales according to predetermined weights to create multi-resolution feature maps. Cross-scale linking, unlike weighted fusion, is a feature alignment technique that aligns feature maps at several sizes to create a multi-resolution feature map for enhanced feature fusion and model performance at many scales.
The fusion process is shown in Figure 3. Multiple convolution and pooling operations are performed on the input image to obtain feature maps of different sizes, and the last three layers are selected to construct the WCFPN, as shown in Equations (1) and (2).
P in = ( p l 1 in , p l 2 in , )
P out = f ( P in )
The multi-scale feature fusion in Equation (1) takes the feature maps from different resolutions and weights them, resulting in a richer and more comprehensive feature representation. Equation (2), on the other hand, cascades feature maps from different resolutions, thus combining feature information from different resolutions in the spatial dimension.
The problem of incomplete and inconsistent feature information is solved by the multi-scale feature fusion method, which can identify and locate targets more accurately. The research feature fusion method uses the fast normalization method, which can adjust the weights of different features by applying a normalization operation on each input feature to fuse feature maps at different scales more effectively, with the formula shown in Equation (3).
O = i ( ω i · X i ) / ( ϵ + j ω j )
where O denotes the output feature mapping; ω i is the weighting factor; X i denotes the input feature mapping, taken as 0.001.
Take D 2 layer output as an example. First, the A 1 layer is convolved 1 × 1 to obtain the feature mapping B 1 . Second, the B 1 layer is fused with the A 2 layer to obtain the B 2 layer, and then the B 1 layer and B 2 layer are fused. Then, fuse both layers to obtain the intermediate layer mapping C. Finally, this layer is fed to D 2 for cross-scale connection and weighting to obtain the D 2 layer. The layer output can be expressed as Equations (4) and (5).
D 2 in = Conv ( ω 1 · B 2 in + ω 2 · Res ( B 1 in ) W 1 + W 2 + ε )
D 2 out = Conv [ ( ω 1 · B 2 in + ω 2 · B 2 td + ω 3 · Res ( B 3 out ) ) / ω 1 + ω 2 + ω 3 + ε ]
where D 2 in is the middle feature of the top-down path layer 2, and D 2 out is the output feature of the bottom-up path layer 2. Res denotes the up-sampling or down-sampling adjustment size, which is used to resize the feature map for feature fusion. All other features are constructed in the same way.

3.2. RWNMS

In target detection, sliding window or region extraction methods are usually used to generate candidate predictor frames, but the same target may be covered by multiple predictor frames, resulting in duplicate and inaccurate target detection results. The role of NMS operation is to select the predictor frame with the highest confidence among these overlapping predictor frames and remove other low-confidence results, and the calculation process is shown in Equation (6).
D 2 out = Conv [ ( ω 1 · B 2 in + ω 2 · B 2 td + ω 3 · Res ( B 3 out ) ) / ω 1 + ω 2 + ω 3 + ε ]
Although NMS can remove overlapping prediction frames, there may be a large amount of overlap in the prediction frames generated on multiple grids or multiple scales of feature maps, which will directly lead to the inability of the NMS algorithm to remove all the redundant prediction frames. To avoid the problem of redundant prediction frames, the study introduces the EIOU loss function, which considers not only the overlap between detection frames but also the similarity and confidence between them, etc.; by giving various detection findings distinct weights, the weighting approach may be utilized to pick and optimize outcomes or instance; while utilizing NMS, some significant targets or detection findings with higher confidence might be assigned larger weights and are hence more likely to be chosen. Additionally, a smoothing strategy is used to further increase the accuracy and robustness of detection by reducing issues like noise and oscillation by weighting the detection results of adjacent frames and averaging them when processing adjacent prediction frames, leading to more accurate and stable detection results; this can successfully boost detection’s resilience and generalization capabilities, better enabling it to handle detection jobs in complicated scenarios. The process of the improved RWI mechanism is shown in Figure 4, and its confidence score is calculated as Equation (7).
c = p 0 × L EIOU .
where L EIOU denotes the intersection ratio of the prediction frame and the true frame obtained based on the EIOU loss function, which can be expressed by Equation (8); p 0 denotes the probability of the presence of the target in the prediction frame, and if it exists then p 0 = 1 , and vice versa p 0 = 0 .
L EIOU = i = 0 s 2 j = 0 B ( L IOU + L dis + L asp ) L dis = ρ 2 ( b , b gt ) ( w c ) 2 + ( h c ) 2 L asp = ρ 2 ( w , w gt ) ( w c ) 2 + ρ 2 ( h , h gt ) ( h c ) 2
where L dis denotes distance loss, L asp denotes the phase loss, and ρ 2 ( b , b gt ) denotes the center point of the two frames b and b gt , w c and h c that denote the length and width of the smallest frame containing A and B, respectively, and the specific parameters are shown in Figure 5.
When there are targets of different scales or different shapes in the target detection, the traditional IOU calculation method is somewhat inaccurate because it only takes into account the ratio of the intersection and the concatenation of the detection frame and true labeled frame, ignoring the shape or size of the frame. To address this issue, EIOU introduces the aspect ratio and the Euclidean distance factor, increasing the variability between IOU values so that larger IOU values are closer to 1 and smaller IOU values are closer to 0, improving the accuracy and robustness of target detection.
In the calculation process, all prediction frames are first arranged in descending order according to the confidence score. Then, starting from the frame with the highest confidence score, all prediction frames are given weights and traversed again, the width and height of the current best frames are updated according to Equation (9), and the current best frame is finally obtained as the result after RWNMS.
Box i = max [ x 1 , y 1 , x 2 , y 2 ] = i = 1 target cls i [ x 1 _ i , y 1 _ i , x 2 _ i , y 2 _ i ] i = 1 target cls i
where box i = max [ ] denotes the box with the highest confidence level, and i = 1 target denotes the prediction box from i = 1 to the target obtained by filtering. After that, the prediction box position is updated by Equation (9) to improve the stability of the prediction results.

4. Experimental Studies

4.1. Data Processing

To construct the data set required for this study, multiple segments of tunnel defective radar data were collected and post-processing operations such as image enhancement and image annotation were performed on these data. The final nearly 5700 images were obtained by LabelImg software labeling to form the data set of this study, and the defect images were classified into five categories such as BM, TK, KD, CS, and YBM, which denote five types of defects, namely, uncompacted, emptying, hollow, water-filled and severely uncompacted, respectively. Table 1 shows the distribution of the five types of defective images in the data set and Figure 6 shows some typical examples of defects.

4.2. Experimental Procedure

4.2.1. Experimental Configuration

Radar detection equipment that emits radar waves and then receives the signals bounced back to obtain information about the parameters shown in Table 2.
The deep learning simulation experiments were built on a Linux system, using Python and PyTorch to build the deep learning framework. The hardware setup shown in Table 3 includes components such as CPU, GPU, memory, and storage.

4.2.2. Evaluation Indicators

The experiments in this study use recall, mAP to evaluate the effectiveness of the model in detecting tunnel defects, which can be defined as Equation (10):
Precision = TP TP + FP Recall = TP TP + FN AP = 0 1 P dR mAP = i = 1 N AP i N mAP [ @ 0.5 , @ 0.95 ] = ( mAP @ 0.5 + + mAP @ 0.95 ) / 10
where TP represents the number of positive examples correctly classified, TN represents the number of negative examples correctly classified, FP represents the number of positive examples incorrectly classified, and FN represents the number of negative examples incorrectly classified.

5. Results

The study used the MFF-YOLO model for tunnel lining defect detection and successfully detected a wide range of defect types and sizes, with some of the results shown in Figure 7.
The loss function curve in Figure 8 was generated using a scatter plot to represent the training data. It is evident from observation that the MFF-YOLO model converges significantly more quickly than other models do, and in the 20th round, all loss functions begin to converge and subsequently begin to stabilize. While the convergence and stabilization of the loss function value suggest that the model has achieved the ideal state, the reduction in the loss function value shows that the model is gradually being optimized during the training phase.
In order to test the effectiveness of each model for each type of defect detection to highlight the importance of each module, the study tallied the accuracy results for different types as shown in Figure 9 and Table 4. 5 sections in Figure 9 represent each of the five types of defects, with different modules added to each section from left to right, where the leftmost is the original model and the rightmost red is the improved MFF-YOLO model. The results show that the improved model proposed in this paper outperforms the original model for all tunnel defect types, which further validates the effectiveness of the model in tunnel defect detection.
To more accurately assess the performance of the MFF-YOLO model proposed in this paper, the study conducted ablation tests and obtained the experimental results shown in Figure 10 and Table 5. The results show that the model MFF-YOLO achieves an accuracy of 89.4%. In addition, through the ablation tests, we also verified that the model has a strong robustness and generalization capability for effective tunnel defect detection under different environments and conditions.
The research in Table 5 shows that after adding WCFPN, the mAP, Precision, Recall and Gflops of the model are significantly improved through deep fusion extraction of feature information. After adding RWNMS to reprocess the prediction box, although the accuracy improvement is not as obvious as WCFPN, it keeps the FPS stable. When the two modules are combined to construct the MFF-YOLO model in this study, the result is a synergistic impact that outperforms the effect of each module acting alone and raises the model’s total performance ceiling.
Figure 11 presents a visualization of the test results of this study on the data set, allowing for a more intuitive understanding of the performance and performance of the model, and thus better model improvement and optimization.
These results show that the improved algorithm proposed in this study has high performance and practicality in tunnel defect detection and can be better adapted to a variety of different types and sizes of defect detection tasks.

6. Conclusions

In this study, we developed an MFF-YOLO model based on multi-scale feature fusion technology. In the neck network, WCFPN is designed based on a weighted fusion strategy to improve the ability to obtain feature information from multiple dimensions. At the predicted end, the prediction mechanism of RWNMS is designed in combination with a weighted smoothing strategy. Finally, the EIOU loss function is improved by considering the aspect ratio and Euclidean distance.
The feasibility of the strategy was evaluated by testing the tunnel defect data set. The result shows that the MFF-YOLO model’s recall rate and accuracy rates were 7.1% and 6.0% higher than those of the YOLOv5 model, reaching 89.5% and 89.4%, respectively.
Compared to conventional image processing approaches, deep learning-based tunnel defect detection algorithms may learn higher-level feature representations from a huge quantity of data, improving the identification and classification of flaws. Additionally, automated and intelligent deep learning-based detection strategies can increase the effectiveness and dependability of detection. In the future, we will continue to explore ways to improve model defect detection performance, including reducing model complexity, improving detection accuracy, and optimizing data set quality.

Author Contributions

Conceptualization, A.Z.; Methodology, A.Z.; Software, B.W. and J.X.; Validation, B.W. and J.X.; Formal analysis, J.X.; Investigation, B.W. and C.M.; Resources, A.Z.; Data curation, C.M.; Writing—original draft, B.W.; Writing—review & editing, A.Z.; Supervision, A.Z.; Project administration, A.Z.; Funding acquisition, A.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Science and Technology Project of Henan Province (222102210135).

Data Availability Statement

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare that they have no competing interest.

References

  1. Wang, J.; Xie, X.; Huang, H. A fuzzy comprehensive evaluation system of mountain tunnel lining based on the fast nondestructive inspection. In Proceedings of the 2011 International Conference on Remote Sensing, Environment and Transportation Engineering, Nanjing, China, 24–26 June 2011; pp. 2832–2834. [Google Scholar]
  2. Ni, Y.; Mao, J.; Fu, Y.; Wang, H.; Zong, H.; Luo, K. Damage Detection and Localization of Bridge Deck Pavement Based on Deep Learning. Sensors 2023, 23, 5138. [Google Scholar] [CrossRef] [PubMed]
  3. Kao, S.-P.; Chang, Y.-C.; Wang, F.-L. Combining the YOLOv4 Deep Learning Model with UAV Imagery Processing Technology in the Extraction and Quantization of Cracks in Bridges. Sensors 2023, 23, 2572. [Google Scholar] [CrossRef]
  4. Santaniello, P.; Russo, P. Bridge Damage Identification Using Deep Neural Networks on Time–Frequency Signals Representation. Sensors 2023, 23, 6152. [Google Scholar] [CrossRef] [PubMed]
  5. Yan, R.; Zhang, R.; Bai, J.; Hao, H.; Guo, W.; Gu, X.; Liu, Q. STMS-YOLOv5: A Lightweight Algorithm for Gear Surface Defect Detection. Sensors 2023, 23, 5992. [Google Scholar] [CrossRef] [PubMed]
  6. Guo, Z.; Wang, C.; Yang, G.; Huang, Z.; Li, G. MSFT-YOLO: Improved YOLOv5 Based on Transformer for Detecting Defects of Steel Surface. Sensors 2022, 22, 3467. [Google Scholar] [CrossRef] [PubMed]
  7. Shaikh, K.; Hussain, I.; Chowdhry, B.S. Wheel Defect Detection Using a Hybrid Deep Learning Approach. Sensors 2023, 23, 6248. [Google Scholar] [CrossRef]
  8. Sjölander, A.; Belloni, V.; Ansell, A.; Nordström, E. Towards automated inspections of tunnels: A review of optical inspections and autonomous assessment of concrete tunnel linings. Sensors 2023, 23, 3189. [Google Scholar] [CrossRef]
  9. Maeda, K.; Takada, S.; Haruyama, T.; Togo, R.; Ogawa, T.; Haseyama, M. Distress Detection in Subway Tunnel Images via Data Augmentation Based on Selective Image Cropping and Patching. Sensors 2022, 22, 8932. [Google Scholar] [CrossRef]
  10. Lei, Y.; Jiang, B.; Su, G.; Zou, Y.; Qi, F.; Li, B.; Jia, F.; Tian, T.; Qu, Q. Application of Air-Coupled Ground Penetrating Radar Based on FK Filtering and BP Migration in High-Speed Railway Tunnel Detection. Sensors 2023, 23, 4343. [Google Scholar] [CrossRef]
  11. Wu, X.; Bao, X.; Shen, J.; Chen, X.; Cui, H. Evaluation of Void Defects behind Tunnel Lining through GPR forward Simulation. Sensors 2022, 22, 9702. [Google Scholar] [CrossRef]
  12. Ali, L.; Alnajjar, F.; Jassmi, H.A.; Gocho, M.; Khan, W.; Serhani, M.A. Performance evaluation of deep CNN-based crack detection and localization techniques for concrete structures. Sensors 2021, 21, 1688. [Google Scholar] [CrossRef]
  13. Wang, A.; Togo, R.; Ogawa, T.; Haseyama, M. Defect detection of subway tunnels using advanced U-Net network. Sensors 2022, 22, 2330. [Google Scholar] [CrossRef]
  14. Li, G.; Ma, B.; He, S.; Ren, X.; Liu, Q. Automatic tunnel crack detection based on u-net and a convolutional neural network with alternately updated clique. Sensors 2020, 20, 717. [Google Scholar] [CrossRef] [Green Version]
  15. Zhu, A.; Chen, S.; Lu, F.; Ma, C.; Zhang, F. Recognition Method of Tunnel Lining Defects Based on Deep Learning. Wirel. Commun. Mob. Comput. 2021, 2021, 9070182. [Google Scholar] [CrossRef]
  16. Zhu, A.; Ma, C.; Chen, S.; Wang, B.; Guo, H. Tunnel Lining Defect Identification Method Based on Small Sample Learning. Wirel. Commun. Mob. Comput. 2022, 2022, 1096467. [Google Scholar] [CrossRef]
  17. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  18. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  19. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  20. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  21. Srinivasu, P.N.; SivaSai, J.G.; Ijaz, M.F.; Bhoi, A.K.; Kim, W.; Kang, J.J. Classification of skin disease using deep learning neural networks with MobileNet V2 and LSTM. Sensors 2021, 21, 2852. [Google Scholar] [CrossRef]
  22. Xu, X.; Zhao, M.; Shi, P.; Ren, R.; He, X.; Wei, X.; Yang, H. Crack detection and comparison study based on faster R-CNN and mask R-CNN. Sensors 2022, 22, 1215. [Google Scholar] [CrossRef] [PubMed]
  23. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  24. Fu, C.Y.; Liu, W.; Ranga, A.; Tyagi, A.; Berg, A.C. Dssd: Deconvolutional single shot detector. arXiv 2017, arXiv:1701.06659. [Google Scholar]
  25. Li, Z.; Zhou, F. FSSD: Feature fusion single shot multibox detector. arXiv 2017, arXiv:1712.00960. [Google Scholar]
  26. Liu, S.; Huang, D. Receptive field block net for accurate and fast object detection. In Proceedings of the European conference on computer vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 385–400. [Google Scholar]
  27. Zhang, S.; Wen, L.; Bian, X.; Lei, Z.; Li, S.Z. Single-shot refinement neural network for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4203–4212. [Google Scholar]
  28. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  29. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  30. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  31. Bochkovskiy, A.; Wang, C.; Liao, H. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  32. Mathias, A.; Dhanalakshmi, S.; Kumar, R. Occlusion aware underwater object tracking using hybrid adaptive deep SORT-YOLOv3 approach. Multimed. Tools Appl. 2022, 81, 44109–44121. [Google Scholar] [CrossRef]
  33. Lai, H.; Chen, L.; Liu, W.; Yan, Z.; Ye, S. STC-YOLO: Small Object Detection Network for Traffic Signs in Complex Environments. Sensors 2023, 23, 5307. [Google Scholar] [CrossRef]
  34. Bao, C.; Cao, J.; Hao, Q.; Cheng, Y.; Ning, Y.; Zhao, T. Dual-YOLO Architecture from Infrared and Visible Images for Object Detection. Sensors 2023, 23, 2934. [Google Scholar] [CrossRef]
  35. Xia, K.; Lv, Z.; Zhou, C.; Gu, G.; Zhao, Z.; Liu, K.; Li, Z. Mixed Receptive Fields Augmented YOLO with Multi-Path Spatial Pyramid Pooling for Steel Surface Defect Detection. Sensors 2023, 23, 5114. [Google Scholar] [CrossRef]
  36. Ruan, Z.; Wang, H.; Cao, J.; Zhang, H. Cross-scale feature fusion connection for a YOLO detector. IET Comput. Vis. 2022, 16, 99–110. [Google Scholar] [CrossRef]
  37. Huang, K.; Li, C.; Zhang, J.; Wang, B. Cascade and fusion: A deep learning approach for camouflaged object sensing. Sensors 2021, 21, 5455. [Google Scholar] [CrossRef]
  38. Mo, L.; Zhu, Y.; Zeng, L. A Multi-Label Based Physical Activity Recognition via Cascade Classifier. Sensors 2023, 23, 2593. [Google Scholar] [CrossRef]
  39. Huang, H.; Tang, X.; Wen, F.; Jin, X. Small object detection method with shallow feature fusion network for chip surface defect detection. Sci. Rep. 2022, 12, 3914. [Google Scholar] [CrossRef] [PubMed]
  40. Xu, Z.; Yang, Y.; Gao, X.; Hu, M. DCFF-MTAD: A Multivariate Time-Series Anomaly Detection Model Based on Dual-Channel Feature Fusion. Sensors 2023, 23, 3910. [Google Scholar] [CrossRef] [PubMed]
  41. Qian, X.; Wang, X.; Yang, S.; Lei, J. LFF-YOLO: A YOLO Algorithm with Lightweight Feature Fusion Network for Multi-Scale Defect Detection. IEEE Access 2022, 10, 130339–130349. [Google Scholar] [CrossRef]
  42. Mao, K.; Jin, R.; Chen, K.; Mao, J.; Dai, G. Trinity-Yolo: High-precision logo detection in the real world. IET Image Process. 2023, 17, 2272–2283. [Google Scholar] [CrossRef]
  43. Wang, J.; Dong, Y.; Zhao, S.; Zhang, Z. A High-Precision Vehicle Detection and Tracking Method Based on the Attention Mechanism. Sensors 2023, 23, 724. [Google Scholar] [CrossRef] [PubMed]
  44. Hu, W.; Cao, L.; Ruan, Q.; Wu, Q. Research on Anomaly Network Detection Based on Self-Attention Mechanism. Sensors 2023, 23, 5059. [Google Scholar] [CrossRef]
  45. Wang, D.; Xiang, S.; Zhou, Y.; Mu, J.; Zhou, H.; Irampaye, R. Multiple-Attention Mechanism Network for Semantic Segmentation. Sensors 2022, 22, 4477. [Google Scholar] [CrossRef]
Figure 1. Ground-Penetrating Radar Vehicle TGRI-GPR.
Figure 1. Ground-Penetrating Radar Vehicle TGRI-GPR.
Sensors 23 06490 g001
Figure 2. Structure of MFF-YOLO. Where the black solid lines indicate the input-output paths and the red solid lines are the additional input-output paths of the improved model, and the black solid lines are used to form a grid map in the detection head part, and the red and blue dashed boxes indicate the location of the prediction box.
Figure 2. Structure of MFF-YOLO. Where the black solid lines indicate the input-output paths and the red solid lines are the additional input-output paths of the improved model, and the black solid lines are used to form a grid map in the detection head part, and the red and blue dashed boxes indicate the location of the prediction box.
Sensors 23 06490 g002
Figure 3. Structure of WCFPN. Where A 1 , A 2 , …, D 1 , D 2 , D 3 represent different upper and lower sampling layers respectively.
Figure 3. Structure of WCFPN. Where A 1 , A 2 , …, D 1 , D 2 , D 3 represent different upper and lower sampling layers respectively.
Sensors 23 06490 g003
Figure 4. Improved RWI module. Where different colors represent different prediction frames.
Figure 4. Improved RWI module. Where different colors represent different prediction frames.
Sensors 23 06490 g004
Figure 5. Some parameters of EIOU loss function.
Figure 5. Some parameters of EIOU loss function.
Sensors 23 06490 g005
Figure 6. Defect sample example. (a) uncompacted, (b) emptying, (c) hollow, (d) water-filled, (e) severely uncompacted.
Figure 6. Defect sample example. (a) uncompacted, (b) emptying, (c) hollow, (d) water-filled, (e) severely uncompacted.
Sensors 23 06490 g006
Figure 7. Detection of different categories of defects. (a) Effect of BM detection, (b) Effect of TK detection, (c) Effect of KD detection, (d) Effect of CS detection, (e) Effect of YBM detection.
Figure 7. Detection of different categories of defects. (a) Effect of BM detection, (b) Effect of TK detection, (c) Effect of KD detection, (d) Effect of CS detection, (e) Effect of YBM detection.
Sensors 23 06490 g007aSensors 23 06490 g007b
Figure 8. MFF-YOLO loss curve.
Figure 8. MFF-YOLO loss curve.
Sensors 23 06490 g008
Figure 9. Comparison of detection effects of different models for different categories.
Figure 9. Comparison of detection effects of different models for different categories.
Sensors 23 06490 g009
Figure 10. Compare of mAP and Recall.
Figure 10. Compare of mAP and Recall.
Sensors 23 06490 g010
Figure 11. Detection result plots. Where (a) is the original defect maps, (b) is the original model detection result maps, (c) is the improved MFF-YOLO model detection result maps, and (d) is the result visualization maps.
Figure 11. Detection result plots. Where (a) is the original defect maps, (b) is the original model detection result maps, (c) is the improved MFF-YOLO model detection result maps, and (d) is the result visualization maps.
Sensors 23 06490 g011
Table 1. Distribution of tunnel defect data sets.
Table 1. Distribution of tunnel defect data sets.
Defect TypeTraining SetValidation SetTest SetLabel
BM1301114760
TK18151271391
KD904782542
CS1136801973
YBM11751111074
Table 2. Radar rover-related parameters.
Table 2. Radar rover-related parameters.
NameConfigure
Equipment ModelTGRI-GPR
Center Frequency200 MHz
Operating Bandwidth100–500 MHz
Depth of detection3 m
Dynamic Range40 dB
Table 3. Experimental hardware configuration.
Table 3. Experimental hardware configuration.
NameConfigure
Operating SystemLinux
Video CardNVIDIA RTX3090
Video Memory24 G
ProcessorIntel€ Core i3-8100
Programming LanguagePython
Deep training frameworkPyTorch
Programming PlatformsPycharm
Table 4. Comparison of experimental results of each model in different defects.
Table 4. Comparison of experimental results of each model in different defects.
TypeAverage Accuracy Rate (%)
Model BMTKKDCSYBM
Yolov595.687.988.476.169.1
SPD90.789.092.678.587.0
MobileNetv389.685.893.167.580.2
WCFPN90.789.092.678.587.1
Alpha-NMS85.887.993.768.589.2
Merge-NMS86.186.795.269.085.5
RWNMS93.786.288.576.981.7
MFF-YOLO93.788.894.187.183.2
Table 5. Results of ablation experiments.
Table 5. Results of ablation experiments.
NamemAP@0.5
(%)
mAP@0.5:0.95
(%)
Precision
(%)
Recall
(%)
GflopsFPS
Fater-RCNN78.848.878.680.288.218.2
RetinaNet81.149.879.781.370.318.3
SSD79.247.075.282.115.265.8
Yolov383.149.180.381.6154.616.4
Yolov583.447.975.482.416.866.6
SPD85.849.180.385.516.071.4
MobileNetv387.750.481.089.415.350.0
WCFPN87.651.682.287.633.152.6
Alpha-NMS85.051.378.683.015.871.4
Merge-NMS84.550.578.381.115.871.4
RWNMS85.449.780.885.916.866.6
MFF-YOLO89.450.882.289.533.350.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, A.; Wang, B.; Xie, J.; Ma, C. MFF-YOLO: An Accurate Model for Detecting Tunnel Defects Based on Multi-Scale Feature Fusion. Sensors 2023, 23, 6490. https://doi.org/10.3390/s23146490

AMA Style

Zhu A, Wang B, Xie J, Ma C. MFF-YOLO: An Accurate Model for Detecting Tunnel Defects Based on Multi-Scale Feature Fusion. Sensors. 2023; 23(14):6490. https://doi.org/10.3390/s23146490

Chicago/Turabian Style

Zhu, Anfu, Bin Wang, Jiaxiao Xie, and Congxiao Ma. 2023. "MFF-YOLO: An Accurate Model for Detecting Tunnel Defects Based on Multi-Scale Feature Fusion" Sensors 23, no. 14: 6490. https://doi.org/10.3390/s23146490

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop