Next Article in Journal
Development of Shoulder Muscle-Assistive Wearable Device for Work in Unstructured Postures
Previous Article in Journal
EfferDeepNet: An Efficient Semantic Segmentation Method for Outdoor Terrain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Bird Nests on Transmission Towers in Aerial Images Based on Improved YOLOv5s

1
Department of Electronic and Electrical Engineering, Wuhan Textile University, Wuhan 430200, China
2
State Key Laboratory of New Textile Materials and Advanced Processing Technologies, Wuhan Textile University, Wuhan 430200, China
3
School of Electrical and Automation, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Machines 2023, 11(2), 257; https://doi.org/10.3390/machines11020257
Submission received: 6 January 2023 / Revised: 31 January 2023 / Accepted: 7 February 2023 / Published: 9 February 2023
(This article belongs to the Section Robotics, Mechatronics and Intelligent Machines)

Abstract

:
To further improve the accuracy of bird nest model detection on transmission towers in aerial images without significantly increasing the model size and to make detection more suitable for edge-end applications, the lightweight model YOLOv5s is improved in this paper. First, the original backbone network is reconfigured using the OSA (One-Shot Aggregation) module in the VOVNet and the CBAM (Convolution Block Attention Module) is embedded into the feature extraction network, which improves the accuracy of the model for small target recognition. Then, the atrous rates and the number of atrous convolutions of the ASPP (Atrous Spatial Pyramid Pooling) module are reduced to effectively decrease the parameters of the ASPP. The ASPP is then embedded into the feature fusion network to enhance the detection of the targets in complex backgrounds, improving the model accuracy. The experiments show that the mAP (mean-Average Precision) of the fusion-improved YOLOv5s model improves from 91.84% to 95.18%, with only a 27.4% increase in model size. Finally, the improved YOLOv5s model is deployed into the Jeston Xavier NX, resulting in a model that runs well and has a substantial increase in accuracy and a speed of 10.2 FPS, which is only 0.7 FPS slower than the original YOLOv5s model.

1. Introduction

According to statistics, bird activities are the third most common cause of high-voltage transmission line faults, after lightning damage and external damage. Furthermore, bird activities on high-voltage transmission towers account for 90% of failures. As a result, an excellent method for detecting bird nests can better prevent electrical failures caused by bird activities [1,2,3].
As most transmission towers are located in fields, and bird nests are mostly built high up in such towers, traditional manual inspection is time-consuming and inefficient [4,5]. With the continuous development of artificial intelligence technology, UAV(Unmanned Aerial Vehicle) inspections are gradually replacing manual inspections. Many intelligent object detection algorithms based on image recognition techniques have been applied to bird nest detection [6,7].
There are two broad categories for detecting bird nests in aerial images in UAV inspections. One category consists of traditional transmission line object detection algorithms based on feature extraction and classification models [8,9,10,11]; however, such algorithms can lack robustness due to the irregular shapes of bird nests and the fact that their color changes as the light changes. The second category is object detection algorithms based on deep learning, including two-stage and one-stage algorithms. The two-stage detection algorithm is executed in two steps: first, the candidate box region is obtained and then the classification of targets within the region is performed; representative algorithms include R-CNN (Rigion-based Convolution Neural Networks), Fast R-CNN, Mask R-CNN, Faster R-CNN, etc. [12,13,14,15]. In ref. [16], the authors proposed an algorithm for bird nest detection based on an improved Faster R-CNN. The effect of complex backgrounds on detecting bird nests was excluded by simultaneously detecting bird nests and transmission towers. The two-stage detection algorithm provides a significant improvement in accuracy over the traditional algorithm, but the detection process needs to be completed in two steps, and the detection speed is slow. The one-stage detection algorithm predicts the class and region of the target directly through the object detection network; representative algorithms include SSD (Single shot multibox detector) and YOLO (You Only Look Once) series [17,18,19,20,21,22]. In ref. [23], the authors replaced the backbone network of the SSD with EfficientNetB7, reducing the model parameters. In ref. [24], the authors combined SSD with HSV (Hue Saturation Value) spatial color filters to improve model accuracy; however, the SSD network did not have the FPN (Feature Pyramid Networks) structure, resulting in a lower accuracy than the YOLO series. As the representative algorithm in the YOLO series, YOLOv3 improved detection accuracy while maintaining the fast detection speed of the one-stage algorithm. In ref. [25], the authors replaced the standard convolution in the backbone of the YOLOv3 network with the depth-separable convolution, thus reducing the network parameters and improving the speed of the detection network; however, the detection of small targets was still inadequate. YOLOv4 builds on the strengths of YOLOv3 and further enhances detection accuracy. In ref. [26], the authors added the Swin Transformer module to the backbone of the YOLOv4 network to address the difficulty of multi-scale detection, but model parameters were too large to be deployed at the edge. In ref. [27], the authors introduced self-attention in the YOLOv4-Tiny network and ported the model to an embedded platform, but it was not accurate enough for target recognition in complex contexts. YOLOv5 has improved the accuracy of YOLOv4 while presenting a lighter version, YOLOv5s. In ref. [28], the authors removed a 20 × 20 prediction head from the prediction network part of YOLOv5s and added a 160 × 160 prediction head to improve the model accuracy, but it did not detect small targets effectively. In ref. [29], the authors replaced the feature fusion network in YOLOv5s with Bi-FPN (Bidirectional Feature Pyramid Network), which improved the model accuracy; however, the structure of Bi-FPN was too complex, leading to a dramatic increase in model size and reducing speed.
YOLOv5 offers higher detection results and detection speed than the YOLOv3 and YOLOv4 algorithms, and its lightweight version, YOLOv5s, achieves 73 FPS on the COCO dataset with a model size of only 27.8 M. However, there are still cases of false detections and misdetections for small targets and targets in complex contexts [30,31]. Complex improvements to the feature fusion layer make the model too large to be applied on embedded devices [32]. To address the above problems, this paper proposes a target detection algorithm for bird nests on transmission towers in aerial images based on improved YOLOv5s. Unlike conventional algorithms, the algorithm in this paper can significantly improve the accuracy without significantly increasing the model size and parameters, thus enabling its successful deployment in the Jetson Xavier NX. The specific aims of the work are as follows: (1) To address the problem of low accuracy of the model, the backbone network of the YOLOv5s model is reconstructed to improve the model accuracy; (2) To address the problem that small targets of distant bird nests are not easy to detect, the attention mechanism is embedded in the feature extraction network to enhance the detection of small targets; (3) To address the problem of poor bird nest detection in complex backgrounds, the ASPP module is improved to enhance the detection of targets in complex backgrounds without significantly increasing the model size.

2. Structure and Features of the YOLOv5 Model

The YOLOv5 target detection algorithm can be divided into four releases: YOLOv5s, YOLOv5m, YOLOv5l, and YOLOv5x, depending on the width and depth of the network. Considering the need to deploy the model to embedded devices at a later stage, the lightweight YOLOv5s was chosen as the base model in this paper. The structure of the YOLOv5s object detection network can be divided into three parts: feature extraction network, feature fusion network, and prediction network, as shown in Figure 1.
The feature extraction network follows the CSPDarkNet in YOLOv4, splitting the residual blocks in DarkNet in the channel dimension direction of the feature map into two parts. Of these, a portion passes typically through the convolution and bottleneck layers, keeping with the structure of Resblock. The other part is spliced with the output of another branch in the direction of the channel of the feature map after a small amount of convolution. This approach reduces computational effort while avoiding problems such as gradient disappearance.
The feature fusion network uses the SPP (Spatial Pyramid Pooling) structure and the PANet (Path Aggregation-Network). The SPP structure solves the problem that the input image size of CNN (Convolutional Neural Networks) must be fixed and can make the input image size unrestricted; the SPP structure is shown in Figure 2.
PANet uses the three feature layers of 80 × 80, 40 × 40, and 20 × 20 extracted by FPN to fuse bottom-up, which can obtain three fused features to improve the effect of target detection at different scales; the structure is shown in Figure 3.
The prediction network implements the prediction of the final information, with three different scales of prediction heads corresponding to three different sizes of fused features. In the model prediction, the prediction network ranks the confidence levels of the N categories of the anchors from highest to lowest, and the threshold setting and redundant borders are removed using a non-maximum suppression algorithm to obtain the category and location information of the predicted target.

3. Recognition of Bird Nests on Transmission Towers Based on Improved YOLOv5s

3.1. Feature Extraction Network Based on OSA Module

To improve the model accuracy, this paper adopts the OSA Block in VOVNet to reconstruct the feature extraction network based on YOLOv5s, replacing the ResBlock in the original feature extraction network with OSA Block.
The OSA Block is an improvement on the Dense Block (whose structure is shown in Figure 4a) in the DenseNet [33,34]. It does away with the fusion method, where each layer aggregates the previous layer to avoid information redundancy. It takes the form of fusing all the previous layers at once in the last layer, so that the input size remains the same and new output channels can be expanded while reducing the number of parameters. The structure of the OSA Block is shown in Figure 4b.

3.2. The Impact of Attentional Mechanism Fusion Studies on the Effectiveness of Small Target Detection

As the feature extraction network continues to deepen and the information extracted at the output becomes progressively more abstract, the initial model is not good enough for small target detection. This paper addresses this problem by incorporating attention mechanisms into the network. The attention mechanism refers to the human behavior of selectively focusing on the important parts of the received information to construct a model that can redistribute the weight of the target information from the irrelevant information in the information received by the network. The SAM (spatial attention module) is primarily designed to capture the correlation between pixel points in the input features; the CAM (channel attention module) aims to enhance the feature channels and amplify the target weights, improving the detection of small targets [35,36,37].
This paper chooses the CBAM which combines the SAM with the CAM [38]. It consists of CAM and SAM sub-modules. It can perform separate attention operations in the channel and spatial dimensions, reducing the model parameters, as shown in Figure 5.
Assuming that the input features are F and the one-dimensional channel convolution in the CAM is M c , then the principle of the CAM is as in Equation (1). Using the CAM output F as the input to the SAM. Assume that the two-dimensional spatial convolution in the spatial attention module is M s . The principle of the spatial attention module is then given by Equation (2).
M s ( F ) = σ ( M L P ( A v g P o o l ( F ) ) + M L P ( M a x P o o l ( F ) ) )
M s ( F ) = σ ( f 7 × 7 ( [ A v g P o o l ( F ) ; M a x P o o l ( F ) ] ) )
In Equations (1) and (2), σ is a sigmoid function and f 7 × 7 is a convolution kernel of size 7 × 7.
The total process of the operation of the CBAM module is shown in Equations (3) and (4).
F = M c ( F ) F
F = M s ( F ) F
The attention mechanism is already widely used in some networks due to its plug-and-play convenience. However, there is no definitive answer as to which part of the network it is better to embed it in. In this paper, two different embedding methods at different locations are designed. One is to integrate the CBAM after the Concat layer in the OSA module, which is in the feature extraction network; the other is to integrate it after the Concat layer in the feature fusion network.
The first embedding method is shown in Figure 6a. We embed the CBAM after the Concat layer of the OSA module. Firstly, the CBAM reconfigures the network structure through the CAM to assign more weight to important semantic information. Secondly, it performs the SAM to compress the number of channels and give better accuracy and a lower error rate.
The second embedding method is shown in Figure 6b. We embed CBAM after the Concat layer in the feature fusion network. The PAN and FPN structures in the feature fusion network allow top-down delivery of semantic information and bottom-up delivery of localization information, and the fusion of deep and shallow information is achieved through four Concat layers. It is possible to assign more weight to the fused feature information if the CBAM module is placed behind the Concat layer.

3.3. Modified ASPP to Enhance Targets Detection in Complex Background

Bird nests on high-voltage transmission towers are mostly in complex contexts, which makes detection much more difficult. In this paper, we introduce the ASPP module to sample the input feature maps according to an atrous convolution of different atrous rates and then fuse the obtained results for average pooling to improve target recognition in complex backgrounds. The structure of the ASPP is shown in Figure 7a. Traditional downsampling increases the field of perception but reduces spatial resolution. Using atrous convolution guarantees resolution while expanding the field of perception. However, the atrous convolution is computed in a chessboard-like format, in which the results obtained at one layer, from an independent set of the previous layer, are not interdependent. If you increase the atrous rate to increase the field of perception of the model, you will lose core information and make the model less accurate. Therefore, this paper improves on the original ASPP structure by turning the four-branch atrous convolution into the three-branch atrous convolution, reducing the number of model parameters. We change the atrous rates from (6, 12, 18, 24) to (3, 5, 7); although this does not enhance the perceptual field as effectively as the original, it avoids information loss as much as possible and thus improves accuracy. The improved structure of the ASPP is shown in Figure 7b.

3.4. The Improved YOLOv5 Algorithm Structure

In summary, the structure of the improved YOLOv5 algorithm is shown in Figure 8. After the input images pass through the feature extraction network, the OSA Block lets the information from the upper to lower layers intersect without losing the core information. After the information intersection, CBAM recalibrates the target feature information and amplifies the target weights to obtain three effective feature layers of different sizes, enhancing the detection of small targets. The feature fusion network is improved with the ASPP module to enhance the perceptual field of each input feature layer. Combining information from different scale feature layers provides the basis for improving the detection accuracy of the model for targets in complex backgrounds.

4. Experimental Results and Analysis

4.1. Data Structure and Processing

In this paper, aerial images of bird nests on high-voltage transmission towers are used as the object of study, and the experimental data are sourced from a power company. In this experiment, 1720 original images are expanded to 2644 images, including a total of 3759 target objects, using data augmentation methods such as flip, light and dark transformation, and Gaussian noise. The ratio of the training set, validation set, and test set is 8:1:1. The final result was 2116 training sets, 264 validation sets, and 264 test sets, and each image was annotated by hand using the LabelImg annotation tool. Figure 9 shows several representative aerial UAVs images. In Figure 9, the bird nest targets are framed in red.

4.2. Experimental Environment

This paper uses a deep learning framework based on the PyTorch 1.6.0 environment with Ubuntu 20.04, python 3.6.10, and CUDA = 11.4, where the training graphics card configuration is an NVIDIA RTXA6000/48G graphics card. The local computer NVIDIA GeForce RTX 3060Ti 8G is used for the trained model test.

4.3. Training Process

During model training, the backbone structure is changed so that no pre-training weights are used for training. To reduce the likelihood of the model falling into a local optimum, the SGD (Stochastic Gradient Descent) optimizer is used. It is trained with 300 epoch rounds and a batch size of 16. The cosine annealing learning rate is used to decay the learning rate of the bias layer to improve the convergence speed of the model.

4.4. Evaluation Indexes

This paper uses common performance evaluation indicators for target detection to compare different algorithms, i.e., Precision, Recall, mAP, FPS (Frames Per Second), Param Size, Total Params, FNR (False Negative Ratio), and FDR (False Detection Ratio).
Precision is shown in Equation (5), where TP (True Positive) represents positive samples predicted by the model to be in a positive class, and FP (False Positive) represents negative samples predicted by the model to be in a positive class. Precision represents the ratio of the actual positive samples to all the positive samples predicted by the model.
Precision = TP TP + FP
Recall is shown in Equation (6), where FN(False Negative) represents the positive samples predicted by the model to be in a negative class. Recall represents the ratio of positive samples predicted by the model to actual positive samples [39].
Recall = TP TP + FN
S = 0 1 p ( r ) d r
The P–R curve plotted with Recall as the horizontal axis and Precision as the vertical axis is integrated to find the area under the curve as mAP, denoted as S, as shown in Equation (7). The P–R curve for the improved model in this paper is shown in Figure 10.
In this paper, we introduced FNR as an index to evaluate the effectiveness of small target detection before and after model improvement. FNR is shown in Equation (8), which represents the ratio of positive samples predicted by the model to be negative samples to all positive samples present.
Furthermore, we introduced FDR as an index to evaluate the effectiveness of the target detection in complex backgrounds before and after model improvement. FDR is shown in Equation (9), which represents the ratio of negative samples predicted as positive by the model to all samples predicted as positive by the model.
FNR = FN FN + TP
FDR = FP FP + TP

4.5. Comparison of Experimental Results

4.5.1. Cross-Direction Comparison of Experimental Results

Table 1 shows the test results of the current mainstream deep learning algorithms on the dataset of this paper [40,41].
The following four conclusions can be drawn from Table 1:
  • Among the five mainstream algorithms of Faster R-CNN, SSD, YOLOv3, YOLOv4, and YOLOv4-Tiny, since Faster R-CNN is a two-stage algorithm, it has the highest mAP at 90.48% but the slowest speed is only 10.1FPS, less than 14% of YOLOv5s speed. YOLOv4-Tiny, with only two prediction heads, has the fastest model detection speed at 175 FPS but the lowest mAP at 80.12, which sacrifices detection accuracy for increased detection speed;
  • The base model chosen in this paper, YOLOv5s, has higher accuracy than all the previous five algorithms and is second only to YOLOv4-Tiny in terms of detection speed and model size, which is why it is chosen as the base algorithm in this paper;
  • The fusion improvement algorithm in this paper improves by 3.34% over the original YOLOv5s, with only a 27.6% increase in model size. Compared to YOLOv5m, which improves network depth and width in YOLOv5, the mAP is improved by 1.41%, the model size is only 46.4% of YOLOv5m, and the detection speed is 4 FPS higher than YOLOv5m, which is eligible for deployment in embedded devices;
  • As Figure 11 shows, the improved fusion algorithm in this paper outperforms the mAP of the latest YOLO series of algorithms such as YOLOXs, YOLOv6s, and YOLOv7-Tiny. Moreover, there is a little difference in terms of speed. This shows that the improved algorithm in this paper is still competitive even when compared to the latest algorithms.

4.5.2. Longitudinal Comparison of Experimental Results

Table 2 shows the algorithms corresponding to the first two improvements in this paper (YOLOv5s-V, YOLOv5s-V-att1, YOLOv5s-V-att2, in that order).
Combining the analysis in Table 2 and Figure 12, the following four conclusions can be drawn:
  • The YOLOv5s-V algorithm only reconstructs the backbone network with the OSA Block in VOVNet. The mAP is increased from 91.84% to 93.08%; FNR is reduced by 0.38%; and the model size is increased by only 2.4MB;
  • The YOLOv5s-V-att1 algorithm is based on YOLOv5s-V, after adding the CBAM attention mechanism to the Concat layer in the feature fusion network. YOLOv5s-V-att1 has a 0.29% increase in mAP compared to YOLOv5s-V without the attention mechanism module, which is not a significant increase. This is because, for feature fusion networks, where the CBAM attention mechanism is added after the feature fusion Concat layer after ResBlock, the feature extraction network loses some of the semantic information;
  • The YOLOv5s-V-att2 algorithm is based on YOLOv5s-V, after adding the CBAM attention mechanism to the Concat layer in the OSA module of the feature extraction network. YOLOv5s-V-att2 has a 1.33% improvement in mAP compared to the original YOLOv5s-V. For the feature extraction network, CBAM performs spatial and channel attention on the fused Concat layer of features in the OSA module, which is good for information retention and weight assignment. Therefore, this paper chose to use the YOLOv5s-V-att2 algorithm, which adds an attention mechanism to the feature extraction network;
  • Compared to the original YOLOv5s, the improved backbone YOLOv5s-V and the YOLOv5s-V-att2 with the attention mechanism have both improved on the aspects of FNR. YOLOv5s-V-att2 improved by 1.08% compared to the original model, which indicates that the improved model has a good improvement in small target detection.

4.5.3. Experimental Comparison of the Improved ASPP Module

Table 3 shows the test results on the original YOLOv5s model with the ASPP module set to the different numbers of atrous convolutions and atrous rates.
The following three conclusions can be drawn from Table 3:
  • From Experiments I, II, and IV, we find that the number of atrous convolutions affects the model parameters but has little effect on mAP; the atrous rate has a greater effect on mAP;
  • From experiments II, IV, V, and VI, we find that the ASPP module with the atrous rate of (3, 5, 7) has the best result, with the mAP 0.06% higher than the original ASPP module with the atrous convolution number of 4 and the model size reduction of 3.76 MB;
  • From experiments V and VII, we find that the mAP is increased by 0.27%, and the model size is increased by 3.76 MB by increasing the number of atrous convolutions by one proportional to the atrous rate along (3, 5, 7). Combining mAP and model size, the ASPP module with many atrous convolutions of 3 and the atrous rate of (3, 5, 7) is chosen in this paper.

4.5.4. Comparison of Results of Ablation Experiments

Table 4 shows the comparison of three improved algorithms for ablation experiments.
Combining the analysis in Table 2 and Figure 12, the following conclusions can be drawn:
  • Improvements to the backbone network resulted in a 1.24% improvement in mAP and an increase in the model size of only 2.4 MB. Adding either the attention mechanism or the Improved ASPP module alone, the attention mechanism works better than the Improved ASPP module because the attention mechanism can better assign weights to detection targets. The reason for the relatively small increase in the mAP of the improved ASPP module is that it duplicates the role of the SPP structure in the backbone network. Therefore, we chose to remove the SPP structure for experimental comparison. It can be seen that after removing the SPP module, YOLOv5s-V-att2 (No SPP) decreased by 0.55% compared to the previous mAP. YOLOv5s-V-Improved ASPP (No SPP) decreased by 0.32% compared to the previous mAP. This suggests that the improved ASPP module does duplicate the role of the SPP structure in the backbone. However, the combined comparison is still the improvement of the attention mechanism that improves the model accuracy more;
  • Both the YOLOv5s-V-improved ASPP and the YOLOv5s-V-att2-improved ASPP have improved on FDR. This represents an improvement in the effectiveness of the improved algorithm in this paper for target detection in complex backgrounds;
  • As Figure 13 shows, the improved model converges faster and has a lower loss value compared to the original model;
  • The three improved ablation experiments resulted in a 3.34% improvement over the original results, with an increase in the model size of only 11.5 M and a reduction in detection speed of 18FPS.
Table 5 shows the comparison of the detection effect of the test set before and after the model improvement. As Table 5 shows.
To further visualize the attention region in the model, this paper presents the fractional heat map visualization analysis of the predicted results from the fusion model. Table 6 shows the test set image thermal comparison chart. As Table 6 shows: the heat-sensing map covers the majority of the detected target area, with the red being the central area where attention can be continuously spread outwards to reduce.

5. Embedded Device Deployment

The embedded device port uses the Jetson Xavier NX, which has 384 CUDA cores, 48 Tensor cores, and 2 NVIDIA engines. It can run multiple modern neural networks in parallel, processing high-resolution data from multiple sensors simultaneously. It is suitable for use in drones, portable medical devices, and other systems. The Jetson Xavier NX is shown in Figure 14a, with real-time data acquisition by calling the hardware camera. The acquisition results are shown in Figure 14b, with an image resolution size of 1280 × 720, a detection speed of 10.2 FPS, and a confidence level of 0.82 in the detection of the bird nests.
Table 7 shows the comparison of Jetson Xavier NX test results. As can be seen from Table 7, the improved YOLOv5s model on the Jetson Xavier NX improved detection accuracy by 3.34% over the original YOLOv5s and reduced detection rate by only 0.7 FPS. The actual test showed the conformity of the results of the requirements for real-time detection of bird nests on high-voltage transmission towers during aerial images by drones.

6. Conclusions

In this paper, we improve the algorithm based on YOLOv5s and reconstruct the backbone network of CSPDarkNet by combining the OSA module in VOVNet, which improves the accuracy of the model without increasing the size of the model or the number of parameters computed and facilitates the deployment of the model. Secondly, we investigate the addition of attention mechanisms in different positions of the model to find the most suitable way to add attention mechanisms to improve the detection of small targets. Finally, we investigate the parameters of the ASPP module and select the most suitable parameters for this model to improve the ASPP for multi-target detection in complex backgrounds. The experimental results show that the accuracy of the model improves from 91.84% to 95.18%, the model size is only 39.85 M, an increase of 27.4% over the initial model. The average detection speed is 55 FPS, and the average detection speed, when ported to embedded devices, is 10.2 FPS, which is suitable for deploying embedded edge computer platforms to meet the requirements of real-time detection for UAV inspection.
The algorithm proposed in this paper is currently used to detect bird nests on transmission towers in aerial images and will subsequently investigate the problem of foreign object faults on aerial images of power transmission lines. Eventually, the algorithm is expected to be extended to the detection and monitoring of other major electrical equipment and to enable cloud collaboration, placing some of the target detection functions on the lightweight end.

Author Contributions

Conceptualization, G.H. and R.W.; methodology, G.H.; software, R.W.; validation, R.W., Q.Y. and S.L.; formal analysis, G.H.; investigation, L.Z.; resources, M.H., S.Y. and L.Q.; data curation, R.W., Q.Y. and L.Q.; writing—original draft preparation, R.W.; writing—review and editing, G.H., R.W. and L.Q.; visualization, M.H.; supervision, S.Y. and L.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key R&D Program of China (No. 2020YFB0905900).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yi, H.; Xiong, Y.; Zhou, G. Overhead transmission line bird damage fault analysis and countermeasures. Power Syst. Technol. 2008, 32, 95–100. [Google Scholar]
  2. Shi, L.; Chen, Y.; Fang, G. Comprehensive identification method of bird’s nest on transmission line. Energy Rep. 2022, 8, 742–753. [Google Scholar] [CrossRef]
  3. Zhang, F.; Wang, W.; Zhao, Y. Automatic diagnosis system of transmission line abnormalities and defects based on UAV. In Proceedings of the 4th International Conference on Applied Robotics for the Power Industry (CARPI), Jinan, China, 11–13 October 2016; pp. 1–5. [Google Scholar]
  4. Wu, X.; Yuan, P.; Peng, Q. Detection of bird nests in overhead catenary system images high-speed rail. Pattern Recognit. 2016, 51, 242–254. [Google Scholar] [CrossRef]
  5. Chao, Y.; Xu, Z.; Yue, Y. Analysis of fault characteristics and preventive measures for bird damage on transmission lines in Hunan. High Volt. Eng. 2016, 42, 3853–3860. [Google Scholar]
  6. Liu, X.; Miao, X.; Jiang, H.; Chen, J. Box-point detector: A diagnosis method for insulator faults in power lines using aerial images and convolutional neural networks. IEEE Trans. Power Deliv. 2021, 36, 3765–3773. [Google Scholar] [CrossRef]
  7. Qiu, Z.; Shi, D.; Kuang, Y. Image Recognition of Harmful Bird Species Related to Transmission Line Outages Based on Deep Transfer Learning. High Volt. Eng. 2021, 47, 3785–3794. [Google Scholar]
  8. Zhou, H.; Yu, G. Research on pedestrian detection technology based on SVM classifier trained by HOG and LTP features. Future Gener. Comput. Syst. 2021, 125, 604–615. [Google Scholar] [CrossRef]
  9. Duan, W.; Tang, P.; Jin, W. Bird’s nest detection on railway contact networks based on HOG features in critical areas. China Railw. 2015, 8, 73–77. [Google Scholar]
  10. Wang, Y.; Shen, Z.; Zhao, H. Fault Diagnosis Method of Photovoltaic Modules Infrared Image Based on Improved CNN-SVM. J. North China Electr. Power Univ. 2022, 1–8. [Google Scholar]
  11. Fang, Q.; Hou, J.; Weng, Y. Risk early warning system for long distance high voltage transmission lines based on GA-BP algorithms. J. Xian Polytech. Univ. 2019, 33, 531–537. [Google Scholar]
  12. Girshick, R. Fast R-CNN. arXiv 2015, arXiv:1504.08083. [Google Scholar]
  13. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  14. He, K.; Gkioxari, G.; Dollar, P. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  15. Ren, S.; He, K. Faster R-CNN: Towards Real-Time Object Detection with Regin Proposal Networks. arXiv 2017, arXiv:1506.01497. [Google Scholar] [CrossRef] [PubMed]
  16. Wang, J.; Luo, H.; Yu, P. Bird’s nest detection in multi-scale of high-voltage tower based on Faster R-CNN. J. Beijing Jiaotong Univ. 2019, 43, 37–43. [Google Scholar]
  17. Liu, W.; Anguelov, D.; Rhan, D. SSD: Single Shot MultiBox Detector; Springer: Cham, Switzerland, 2016. [Google Scholar]
  18. Redmon, J.; Divvala, S.; Girshick, R. You Only Look Once: Unified, Real-Time Object Detection. arXiv 2016, arXiv:1506.02640. [Google Scholar]
  19. Redmon, J.; Farrhadi, A. YOLO9000: Better, Faster, stronger. In Proceedings of the IEEE Conference on Computer Vision & Patt-ern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  20. Redmon, J.; Farrhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  21. Bochkovskiy, A.; Wang, C.; Liao, H. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  22. YOLOv5. Available online: https://github.com/ultralytics/yolov5 (accessed on 21 May 2022).
  23. Satheeswari, D.; Shanmugam, L.; Swaroopan, N. Recognition of Bird’s Nest in High Voltage Power Line using SSD. In Proceedings of the 1st International Conference on Electrical, Electronics, Information and Communication Technologies, ICEEICT, Trichy, India, 16–18 February 2022. [Google Scholar]
  24. Jiang, H.; Huang, W. Detection of Bird Nests on Power Line Patrol Using Single Shot Detector. In Proceedings of the 2019 Chinese Automation Congress, Hangzhou, China, 22–24 November 2019. [Google Scholar]
  25. Yang, X.; He, P.; Chen, L.; Li, J. Bird’s Nest Detection on Lightweight YOLOv3 Transmission Line Based on Deep Separable Convolution. Smart Power 2021, 49, 88–95. [Google Scholar]
  26. Zhang, Z.; He, G. Recognition of Bird Nests on Power Transmission Lines in Aerial Images Based on Improved YOLOv4. Front. Energy Res. 2022, 10, 870253. [Google Scholar] [CrossRef]
  27. Han, G.; He, M.; Zhao, F. Insulator detection and damage identification based on improved lightweight YOLOv4 network. Energy Rep. 2021, 7, 187–197. [Google Scholar] [CrossRef]
  28. Ying, Z.; Lin, Z.; Wu, Z.; Liang, K.; Hu, X. A modified-YOLOv5s model for detection of wire braided hose defects. Measurement 2022, 190, 110683. [Google Scholar] [CrossRef]
  29. Han, G.; He, M.; Gao, M.; Yu, J.; Liu, K.; Qin, L. Insulator Breakage Detection Based on Improved YOLOv5. Sustainability 2022, 14, 6066. [Google Scholar] [CrossRef]
  30. Zhang, Z.; Zhang, B.; Lan, Z. FINet: An Insulator Dataset and Detection Benchmark Based on Synthetic Fog and Improved YOLOv5. IEEE Trans. Instrum. Meas. 2022, 71, 99. [Google Scholar] [CrossRef]
  31. Wang, H.; Xu, Y.; He, Y. YOLOv5-Fog: A Multiobjective Visual Detection Algorithm for Fog Driving Scenes Based on Improved YOLOv5. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  32. Deng, F.; Xie, Z.; Mao, W. Research on edge intelligent recognition method oriented to transmission line insulator fault detection. Int. J. Electr. Power Energy Syst. 2022, 139, 108054. [Google Scholar] [CrossRef]
  33. Gao, H.; Zhuang, L. Densely Connecter Convolution Networks. arXiv 2016, arXiv:1608.06993. [Google Scholar]
  34. Bai, H.; Tang, B.; Cheng, T.; Liu, H. High impedance fault detection method in distribution network based on improved Emanuel model and DenseNet. Energy Rep. 2022, 8, 982–987. [Google Scholar] [CrossRef]
  35. Ren, H.; Wang, X. The overview of attentional mechanisms. J. Comput. Appl. 2021, 41, 1–6. [Google Scholar]
  36. Ye, T.; Zhang, J.; Li, Y.; Zhang, X.; Zhao, Z.; Li, Z. CT-Net: An Efficient Network for Low-Altitude Object Detection Based on Convolution and Transformer. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  37. Wang, L.; Cao, Y.; Wang, S.; Song, X.; Zhang, S.; Zhang, J.; Niu, J. Investigation into Recognition Algorithm of Helmet Violation Based on YOLOv5-CBAM-DCN. IEEE Access 2022, 10, 60622–60632. [Google Scholar] [CrossRef]
  38. Woo, S.; Park, J. CBAM: Convolution Block Attention Module. arXiv 2018, arXiv:1807.06521. [Google Scholar]
  39. Jabir, B.; Falih, N. Deep learning-based decision support system for weeds detection in wheat fields. Int. J. Electr. Comput. Eng. 2022, 12, 816. [Google Scholar] [CrossRef]
  40. Li, C.; Li, L.; Jiang, H. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  41. Yao, C.; Bochkovskiy, A.; Liao, H. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
Figure 1. The structure of the original YOLOv5s.
Figure 1. The structure of the original YOLOv5s.
Machines 11 00257 g001
Figure 2. The structure of the SPP.
Figure 2. The structure of the SPP.
Machines 11 00257 g002
Figure 3. The structure of the PANet.
Figure 3. The structure of the PANet.
Machines 11 00257 g003
Figure 4. (a) Structure of the Dense Block; (b) structure of the OSA Block.
Figure 4. (a) Structure of the Dense Block; (b) structure of the OSA Block.
Machines 11 00257 g004
Figure 5. The structure of the CBAM.
Figure 5. The structure of the CBAM.
Machines 11 00257 g005
Figure 6. (a) CBAM fusion in feature extraction network; (b) CBAM fusion in feature fusion network.
Figure 6. (a) CBAM fusion in feature extraction network; (b) CBAM fusion in feature fusion network.
Machines 11 00257 g006
Figure 7. (a) Structure of the original ASPP; (b) structure of the improved ASPP.
Figure 7. (a) Structure of the original ASPP; (b) structure of the improved ASPP.
Machines 11 00257 g007
Figure 8. The structure of the improved YOLOv5s.
Figure 8. The structure of the improved YOLOv5s.
Machines 11 00257 g008
Figure 9. (a) Single-target bird nest; (b) multi-target bird nests; (c) bird nest targets in complex backgrounds; (d) bird nests as small targets.
Figure 9. (a) Single-target bird nest; (b) multi-target bird nests; (c) bird nest targets in complex backgrounds; (d) bird nests as small targets.
Machines 11 00257 g009
Figure 10. The P–R curves for the improved model.
Figure 10. The P–R curves for the improved model.
Machines 11 00257 g010
Figure 11. The FPS–mAP scatter plots.
Figure 11. The FPS–mAP scatter plots.
Machines 11 00257 g011
Figure 12. (a) Training loss curves for the models are shown in Table 2; (b) validation loss curves for the models are shown in Table 2.
Figure 12. (a) Training loss curves for the models are shown in Table 2; (b) validation loss curves for the models are shown in Table 2.
Machines 11 00257 g012
Figure 13. (a) Training loss curves for the models are shown in Table 4; (b) validation loss curves for the models are shown in Table 4.
Figure 13. (a) Training loss curves for the models are shown in Table 4; (b) validation loss curves for the models are shown in Table 4.
Machines 11 00257 g013
Figure 14. (a) Structure of Jetson Xavier NX; (b) real-time monitoring and collection results.
Figure 14. (a) Structure of Jetson Xavier NX; (b) real-time monitoring and collection results.
Machines 11 00257 g014
Table 1. Horizontal comparison of different algorithms.
Table 1. Horizontal comparison of different algorithms.
ModelsmAP/%Speed/FPSModel Size/MB
Faster R-CNN89.6514113
SSD90.48101100.27
YOLOv389.1055236
YOLOv490.3544245
YOLOv4-Tiny80.1217523
Original YOLOv5s91.847327.8
YOLOv5m93.775181.54
YOLOXs94.527635.9
YOLOv6s93.96517.2
YOLOv7-Tiny91.968824.2
Improved YOLOv5s95.185537.85
Table 2. Test results for the fusion of reconstructed backbone and attention mechanisms.
Table 2. Test results for the fusion of reconstructed backbone and attention mechanisms.
ModelsmAP/%Speed/FPSFNR/%Model Size/MB
YOLOv5s91.84732.8427.8
YOLOv5s-V93.08652.4630.2
YOLOv5s-V-att193.37612.1131.28
YOLOv5s-V-att294.41611.7631.58
Table 3. ASPP module improvement test experiment.
Table 3. ASPP module improvement test experiment.
ExperimentsThe Number of Atrous ConvolutionsAtrous RatesmAP/%Model Size/MB
I4(6, 12, 18, 24)92.9538.11
II3(6, 12, 18)91.7334.35
III5(6, 12, 18, 24, 30)92.5841.86
IV3(4, 6, 8)92.9234.35
V3(3, 5, 7)93.0134.35
VI3(2, 4, 6)92.6534.35
VII4(3, 5, 7, 9)93.2838.11
Table 4. Experimental results of ablation with three improved algorithms.
Table 4. Experimental results of ablation with three improved algorithms.
ModelsmAP/%Speed/FPSFDR/%Model Size/MB
YOLOv5s91.84736.5627.8
YOLOv5s-V93.08656.4430.2
YOLOv5s-V-att294.41615.931.58
YOLOv5s-V-att2(No SPP)93.86636.429.58
YOLOvs5-V-Improved ASPP93.97605.4238.47
YOLOvs5-V-Improved ASPP(No SPP)93.65615.6436.47
YOLOv5s-V-att2-Improved ASPP95.18555.1739.85
Table 5. Test set image detection results. (a) the improved model reduces false detections; (b) the improved model enhances the detection of small targets; (c,d) the improved model enhances the detection of multiple targets in complex backgrounds.
Table 5. Test set image detection results. (a) the improved model reduces false detections; (b) the improved model enhances the detection of small targets; (c,d) the improved model enhances the detection of multiple targets in complex backgrounds.
Original AlgorithmImproved Algorithm
(a) Machines 11 00257 i001 Machines 11 00257 i002
(b) Machines 11 00257 i003 Machines 11 00257 i004
(c) Machines 11 00257 i005 Machines 11 00257 i006
(d) Machines 11 00257 i007 Machines 11 00257 i008
Table 6. Test set image thermal comparison chart. (a) the improved model detects targets that the initial model sensed but could not detect; (b) the improved model provides a significant improvement in the detection of small targets; (c) the improved model addresses false detection in complex backgrounds; (d) the improved model enhances multi-target detection in complex backgrounds.
Table 6. Test set image thermal comparison chart. (a) the improved model detects targets that the initial model sensed but could not detect; (b) the improved model provides a significant improvement in the detection of small targets; (c) the improved model addresses false detection in complex backgrounds; (d) the improved model enhances multi-target detection in complex backgrounds.
Original AlgorithmImproved Algorithm
(a) Machines 11 00257 i009 Machines 11 00257 i010
(b) Machines 11 00257 i011 Machines 11 00257 i012
(c) Machines 11 00257 i013 Machines 11 00257 i014
(d) Machines 11 00257 i015 Machines 11 00257 i016
Table 7. The comparison of Jetson Xavier NX test results.
Table 7. The comparison of Jetson Xavier NX test results.
ModelsmAP/%Speed/FPS
Original YOLOv5s91.8410.9
Improved YOLOv5s95.1810.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, G.; Wang, R.; Yuan, Q.; Li, S.; Zhao, L.; He, M.; Yang, S.; Qin, L. Detection of Bird Nests on Transmission Towers in Aerial Images Based on Improved YOLOv5s. Machines 2023, 11, 257. https://doi.org/10.3390/machines11020257

AMA Style

Han G, Wang R, Yuan Q, Li S, Zhao L, He M, Yang S, Qin L. Detection of Bird Nests on Transmission Towers in Aerial Images Based on Improved YOLOv5s. Machines. 2023; 11(2):257. https://doi.org/10.3390/machines11020257

Chicago/Turabian Style

Han, Gujing, Ruijie Wang, Qiwei Yuan, Saidian Li, Liu Zhao, Min He, Shiqi Yang, and Liang Qin. 2023. "Detection of Bird Nests on Transmission Towers in Aerial Images Based on Improved YOLOv5s" Machines 11, no. 2: 257. https://doi.org/10.3390/machines11020257

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop