Next Article in Journal
A Ship Detection and Imagery Scheme for Airborne Single-Channel SAR in Coastal Regions
Next Article in Special Issue
YOLO-SD: Small Ship Detection in SAR Images by Multi-Scale Convolution and Feature Transformer Module
Previous Article in Journal
Large-Scale Monitoring of Glacier Surges by Integrating High-Temporal- and -Spatial-Resolution Satellite Observations: A Case Study in the Karakoram
Previous Article in Special Issue
Micro-Doppler Curves Extraction of Space Target Based on Modified Synchro-Reassigning Transform and Ridge Segment Linking
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SEAN: A Simple and Efficient Attention Network for Aircraft Detection in SAR Images

1
Tianjin Key Lab for Advanced Signal Processing, Civil Aviation University of China, Tianjin 300300, China
2
Engineering Techniques Training Center, Civil Aviation University of China, Tianjin 300300, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(18), 4669; https://doi.org/10.3390/rs14184669
Submission received: 11 July 2022 / Revised: 6 September 2022 / Accepted: 15 September 2022 / Published: 19 September 2022
(This article belongs to the Special Issue Microwave Remote Sensing for Object Detection)

Abstract

:
Due to the unique imaging mechanism of synthetic aperture radar (SAR), which leads to a discrete state of aircraft targets in images, its detection performance is vulnerable to the influence of complex ground objects. Although existing deep learning detection algorithms show good performance, they generally use a feature pyramid neck design and large backbone network, which reduces the detection efficiency to some extent. To address these problems, we propose a simple and efficient attention network (SEAN) in this paper, which takes YOLOv5s as the baseline. First, we shallow the depth of the backbone network and introduce a structural re-parameterization technique to increase the feature extraction capability of the backbone. Second, the neck architecture is designed by using a residual dilated module (RDM), a low-level semantic enhancement module (LSEM), and a localization attention module (LAM), substantially reducing the number of parameters and computation of the network. The results on the Gaofen-3 aircraft target dataset show that this method achieves 97.7% AP at a speed of 83.3 FPS on a Tesla M60, exceeding YOLOv5s by 1.3% AP and 8.7 FPS with 40.51% of the parameters and 86.25% of the FLOPs.

1. Introduction

Synthetic aperture radar (SAR) is an active microwave imaging sensor with all-day and all-weather capabilities; it is widely used in natural disaster monitoring, military reconnaissance, and urban planning [1,2]. As a typical target, aircraft have essential value in both military and civilian fields. Therefore, SAR aircraft interpretation has always been one of the research hotspots. With the continuous improvement of SAR imaging resolution, there are higher requirements for the accuracy and speed of aircraft target detection.
Traditional SAR target detection methods mainly include the constant false alarm rate (CFAR) algorithm, based on the statistical distribution of background clutter [3,4], and the algorithm of manually extracted image features [5,6]. The CFAR algorithm is primarily influenced by the statistical characteristics of background clutter, and it is difficult to accurately model the scene of complex objects, which generates more false alarms. The algorithm for manual extraction of image features divides detection into two separate processes: feature extraction and target classification. However, the process of manually extracting features is complex and highly dependent on parameter settings. Thus robustness and generalization are difficult to guarantee.
With the rapid development of deep learning methods and the increase of high-resolution SAR data, the emergence of convolutional neural network (CNNs) has brought many breakthroughs in tasks such as SAR image segmentation [7], land-cover classification [8], and target detection [9]. Data-driven models have good robustness and generalization. At the same time, this method can actively extract high-level features, avoiding the complicated work of manual feature selection in traditional algorithms. For the SAR aircraft detection task, He et al. [10] explored the positional relationship between the two discrete components of the head and tail of the aircraft target, detected the two components respectively, and then used the KNN algorithm to match the discrete components. Guo et al. [11] first extracted airport areas and clustered these regions to find suspected aircraft targets, then enhanced the scattering information of these regions of interest and sent them to the network for training. Zhao et al. [12] introduced the attention mechanism and dilated convolution to improve the design of the neck of the feature pyramid network (FPN) [13] to improve the detection accuracy. Kang et al. [9] used an attention mechanism at the neck of the network based on the FCOS [14] network. They introduced a priori information of strong scattering points to the output of multiple-scale detection heads, which can effectively filter out false alarms.
Although the above deep learning-based SAR aircraft target detection algorithms have high detection accuracy, they usually use large backbone networks [9,11,12,15,16] such as ResNet 50/101 [17] or complex feature pyramid [9,11,12,14,18,19,20,21,22] neck design. These operations are complicated, and the expansibility is insufficient. An overly large network model not only has a high training time cost but also reduces the efficiency of aircraft target detection during inference due to excessive parameters and calculations. Compared with optical large-scale benchmark datasets such as Microsoft COCO [23] (328 K images with 80 categories), the SAR dataset is small, and the semantic information is relatively simple, such as SSDD [24] (1.16 K images with one category). Therefore, if a large network model is used in the SAR target detection task, it is necessary to load a pre-training model for good parameter initialization. Otherwise, there is severe over-fitting. In addition, for the problem of less SAR label data, there are currently studies such as semi-supervised learning [25] or generative adversarial [26] to expand the dataset to solve the problem effectively. Considering the unique imaging mechanism of SAR images, aircraft targets are in a discrete state in the image, and their detection performance is easily affected by multi-faceted structures such as covered bridges and airport buildings. To address the above problems, we propose a simple and efficient attention network (SEAN) for aircraft target detection in SAR images in this paper. First, a shallow and efficient backbone network is designed by introducing a structural re-parameterization technique [27]. Second, combining the unique scattering mechanism of SAR images with the a priori information of aircraft, we design a simple network neck with no complex lateral connections and multi-scale structure. This structure utilizes the dilated convolution of residual connections combined with the shallow semantic information enhanced by the attention mechanism and can perform aircraft target detection well with only one scale of feature resolution.
The main contributions of our work in this paper can be summarized as follows:
(1)
The proposed SEAN is a SAR aircraft target detection network with a simple structure, high accuracy, and high speed. Compared with the typical target detection algorithms from recent years, SEAN has apparent advantages in detection accuracy and speed on the Gaofen-3 dataset.
(2)
An appropriate network size is selected to balance the detection accuracy and speed. Then, the backbone network’s depth is explored, proving that the C4 feature of the backbone network is more suitable for aircraft target detection. Furthermore, this paper uses a structural re-parameterization technique on the shallowed backbone to effectively enhance the feature extraction capability.
(3)
A simple and efficient neck of the network is designed, discarding the complex feature pyramid network design. It mainly consists of three modules. One is a residual dilated module (RDM) that integrates the multi-scale receptive field. The second is a low-level semantic enhancement module (LSEM) that enhances the scattered information of SAR images. Furthermore, the third is a location attention module (LAM) that refines the multi-feature information after fusion.

2. Related Work

2.1. Deep Learning-Based Object Detection Algorithm

Object detection algorithms based on deep learning can be divided into two-stage [28,29] and one-stage [14,30,31] according to the number of detection stages. A two-stage detection algorithm has two detection heads. The first stage detects the candidate regions where there may be objects, and the second stage predicts the category and position of the object in the candidate regions. The two-stage detection algorithm represented by Faster-RCNN [29] usually has high detection accuracy. However, more complex structures have poorer detection speed than one-stage methods. The one-stage detection algorithm represented by YOLO [32] directly predicts the category and position of the object. This end-to-end method usually has faster detection speed and good detection accuracy. To obtain better feature extraction ability on large benchmark datasets such as ImageNet, a large backbone network such as ResNet101 is usually used. Recently, Vision Transformer (ViT) [33] as a backbone network has shown better feature extraction ability than CNN, which performs well on various visual downstream tasks [34]. However, this work shows that the performance of ViT exceeds that of CNN on the premise that there are hundreds of millions of data for pre-training. Otherwise, ViT’s performance is not as good as CNN on datasets below a million. Therefore, in SAR aircraft detection with a small amount of data, CNN-based detection algorithms are mainly used for research [10,11,12]. Luo et al. [20] showed that the CSPDarknet backbone of the YOLO series is less affected by background interference, and it is more suitable for SAR aircraft target detection tasks than backbones such as ResNet.
For a long time, the YOLO series of algorithms [30,31,32,35,36,37] have always pursued the best balance between accuracy and speed, and the method in this paper is an improvement based on YOLOv5 6.0 [37]. YOLOv5 provides a total of five model sizes of n, s, m, l, and x according to the different widths and depths of the network. As shown in Figure 1, the network architecture of YOLOv5 6.0 can be divided into three parts: backbone, neck, and head. The backbone is an optimized CSPDarknet that has five stages for feature extraction and generates five scale features, and then sends the C3, C4, and C5 feature maps output by the last three stages to the neck. The neck is the optimized PANet [38], which has two branches for top-down and bottom-up multi-scale feature fusion. The head follows the coupled head of YOLOv3 [30] and detects objects of large, medium, and small scales based on nine preset anchor boxes. The loss function of the algorithm is composed of BCE as classification loss and objectness loss, and CIoU [39] as regression loss for bounding box prediction.

2.2. Structure of the Detector Neck

In the typical object detection algorithms in recent years, it is indispensable to use the feature pyramid network (FPN) [13] as the neck of the network. Early CNNs [32] only used the C5 feature for detection, and its performance for small target detection was generally not high. Since then, the proposed method of top-down delivery of higher-order semantic information by FPN has effectively solved the low performance of small target detection. FPN usually brings two benefits: (1) performs multi-scale feature fusion, and (2) assigns objects of different scales to different receptive field features for detection. However, people often attribute the effectiveness of FPN to its ability to perform multi-scale feature fusion. With the success of FPN, some effective improvement works have appeared one after another. PANet [38] adds a bottom-up line based on FPN, which can pass up the low-level semantic information for spatial localization for fusion. Subsequently, there have been some works on complex laterally connected feature pyramids, such as BiFPN [40] and NAS-FPN [41]. These aim to obtain good detection accuracy through better multi-scale feature fusion. Recent SAR aircraft target detection algorithms [9,11,12,14,18,19,20,21,22] use an FPN as the neck. They usually introduce attention mechanisms [9,11,12,18] in the lateral connections to enhance the learning of channel and spatial information.
However, these FPN methods impose a large memory and computational burden, resulting in low efficiency for object detection. A recent work, YOLOF [42] showed that the significant effect brought by FPN mainly solves the optimization problem of multi-scale object detection in a divide-and-conquer way and does not rely on the part of multiscale feature fusion. Moreover, it proposes a simple neck design with a label-matching mechanism that can have excellent detection accuracy even with only the top C5 feature, but its detection performance for small objects is poor compared to that of YOLOv4 [35]. Such a finding is not accidental, and some recent work on object detection in ViT has also shown the non-essentiality of complex lateral connections. ViTDet [34] showed that operation without lateral connections achieved good performance only using the topmost feature by simply constructing a pyramid structure by downsampling and up-sampling. In addition, AdaMixer [43] designed an adaptive sampling decoder to replace the FPN that can converge quickly and has good detection performance. It is worth mentioning that when AdaMixer did the ablation experiment of the backbone, it was found that the accuracy of ResNet only using the C4 feature for detection was better than that of the C3 and C5 features, which is similar to the conclusion of the work on the shallowed backbone in this paper.

2.3. Attention Mechanism

In recent years, attention mechanisms have emerged to allow computer vision systems to mimic the human visual system and find salient regions in complex scenes naturally and efficiently. The attention mechanism in computer vision can be regarded as the process of dynamically adjusting the weights based on the features of the input image [44]. This mechanism is used as a plug-and-play module, which can exchange high performance with a small amount of computational overhead and is now widely used in various computer vision tasks.
The detection network usually adds attention mechanisms such as SE [45] and CBAM [46] to enhance performance. However, SE only considers the connection between internal channel information and ignores the importance of spatial information. At the same time, CBAM introduces the local pooling operation of spatial information to the channel, but it suffers from the inability to obtain a large range of dependent information. The proposed coordinate attention mechanism (CoordAtt) [47] alleviates the above problems by embedding large-scale location information into channel attention. The structure of CoordAtt is shown in Figure 2 and consists of two main steps: coordinate information embedding and coordinate attention generation. In the first step, the input feature map X is first aggregated into two separate orientation-aware feature maps along the horizontal and vertical directions through two one-dimensional global pooling operations. In the second step, the two feature maps are first spliced with embedded specific orientation information, and then, a 1 × 1 convolution and sigmoid function is used to output an intermediate feature map encoded with horizontal and vertical spatial information. The intermediate feature map is then split into two separate vectors along the spatial dimension, the channels are raised to the same dimension as the input feature map through two 1 × 1 convolutions, and the attention maps are output by the sigmoid function as g h and g w . Each attention map captures long-range dependencies of input feature maps along one spatial direction, and location information can thus be preserved in the generated attention map. For example, Formula (1) calculates the weight of y c of the cth channel of the output:
y c ( i , j ) = x c ( i , j ) × g c h ( i ) × g c w ( j )
The representation of attention regions can be emphasized by applying both attention maps to the input feature map by multiplication.

3. Methodology

3.1. Overview of the Architecture of the Proposed SEAN

In order to detect aircraft accurately and quickly in SAR images with complex background interference, we propose a simple and efficient attention network (SEAN) architecture, as shown in Figure 3. Considering the trade-off between speed and accuracy, this paper selects YOLOv5s version 6.0 [37] as our baseline and makes a series of improvements for the SAR aircraft target detection task. The detailed network configuration of SEAN is shown in Table 1. In the table, n represents the stacked times of the module, and arguments represent the parameter information of the module, including input channel, output channel, kernel size, stride, and padding.
The backbone’s primary function is to obtain good feature extraction ability. In this paper, the depth of the backbone network is preliminarily explored for the SAR aircraft target detection task, and structural re-parameterization technology [27,48] is introduced to enhance its feature extraction ability. For the neck, different from the usual complex lateral connection design to achieve multi-scale detection and multi-feature fusion, this paper designs a simple neck consisting of three parts: a residual dilated module (RDM), a low-level semantic enhancement Module (LSEM), and a localization attention module (LAM). The following is a detailed introduction to the method in this paper.

3.2. Optimization of Backbone Network

SAR images are presented as single-channel grayscale images, which contain far less semantic information than the optical benchmark dataset. Moreover, aircraft are mainly displayed in the images as discrete strong scattering points, which are concentrated in the low-level semantic information. From the perspective of the receptive field, aircraft are usually a small target [49] in SAR images, and it is not necessary to use a deeper backbone to obtain a larger receptive field. In accordance with the above data characteristics, this section explores the network depth settings and introduces a structural re-parameterization technique to obtain better feature extraction capabilities.

3.2.1. Shallow Backbone

The backbone usually has five stages for hierarchical feature extraction. First, the resolution of the feature map of each higher stage is double downsampled, and the number of channels is doubled. Then, the last three stages of the backbone output C3, C4, and C5 features, which are used for multi-scale object detection. Since this paper does not use the complex FPN neck for target detection, it is necessary to explore at which stage the output feature map is more suitable for the aircraft detection task. Therefore, we make a preliminary exploration of the depth of the backbone. Note that the depth here refers to the number of stages selected. We use the C3, C4, and C5 output features with only one detection head for the aircraft detection task at a single resolution. We find that only the C4 feature map has good detection results, which may be attributed to the higher resolution of the C4 feature map being more suitable for medium and small aircraft detection tasks. Therefore, this paper uses the backbone without the fifth stage as the basis for subsequent work.

3.2.2. Backbone Re-Parameterization

Before each stage of the backbone performs feature extraction, the input feature map is downsampled to reduce the computational load of the subsequent network. However, downsampling reduces the resolution and inevitably loses some detailed information, which is not conducive to positioning aircraft targets. As shown in Figure 4a, the basic Conv block is downsampled by convolution with kernel 3 and stride 2, followed by the BN layer and SiLU activation function to maintain better gradient transfer. To better retain feature information during downsampling, this paper introduces a structural re-parameterization technique of RepVGG [27] to optimize the Conv block. The core idea of the structural re-parameterization technique is that the network realizes the equivalent conversion of the structure through the equivalent conversion of parameters. Specifically, a set of branches with a convolution kernel size of 1 are added to the original Conv block during model training. The RepConv block is formed to increase the diversity of downsampling, and the parameters of the added branch are fused into the Conv block using the re-parameterization technique during inference so that the structure of the inference is the same as that of the Conv block, which enhances the capability of the model without affecting the detection speed. Convolution in deep learning is a cross-correlation operation. Since this operation is linear, the parameters of the convolution kernel increase with additivity. Figure 4b shows the simplified RepConv re-parameterization process. During inference, firstly, the BN layer is just a simple linear mapping, and its parameters can be directly added to the convolution kernel to improve the inference speed. It can be assumed that the number of channels in the input feature map X is 1, and the number of channels after downsampling by RepConv is 2. Then, use A R 1 × 2 × 3 × 3 and B R 1 × 2 × 1 × 1 to represent the weight parameters of the 3 × 3 and 1 × 1 convolutional layers, respectively. For a 1 × 1 convolution kernel, it can be equated to a special 3 × 3 convolution kernel with only non-zero values at the center. So that the weights of the convolution layers of the two branches can be summed up by channels and the summed weights can be denoted as C R 1 × 2 × 3 × 3 , this paper uses “*” to represent the convolution operation; then, this process can be represented by Formula (2):
X A + X B = X ( A + B ) A + B = C X A + X B = X C

3.3. Neck Design of SEAN

FPN [13] is proposed to use a divide-and-conquer approach for multi-scale target detection, which alleviates the complex detection problem of small target objects. However, this complex lateral connection approach brings a significant memory and computation burden, which reduces detection efficiency. In this section, we wish to explore a simple and efficient network of necks to ensure accurate and fast SAR aircraft detection. We use the coordinate attention mechanism (CoordAtt) [47] as the base component of the neck to better capture the spatial localization information of the aircraft. In the following subsections, the three modules that make up the neck are described individually.

3.3.1. Residual Dilated Module (RDM)

Considering that too much downsampling brings a larger receptive field, it also loses many details required for small target detection, and we hope to perform the detection task well on a feature map with only one resolution size; the dilated encoder approach [42] is used to alleviate these problems. We use dilated convolution after the optimized backbone to increase the receptive field while maintaining the original feature map size. As shown in Figure 5, we use the Conv block to design a residual dilated module (RDM), which is formed by stacking two residual blocks with different expansion rates. The C4 feature map is the input to this module, which first enters the 1 × 1 convolution to reduce the channel dimension. Then, we use the dilated convolution to expand the receptive field, and we finally revert to the original channel number and fuse it with the input feature with the smaller receptive field. Figure 6 indicates the size of the aircraft targets covered by the receptive field in the feature map: (a) indicates that the receptive field of the C4 feature can cover most of the small- and medium-sized targets; (b) indicates the receptive field after dilated convolution can cover large- and medium-sized targets; and (c) indicates that the features of the two receptive fields can be fused to cover almost all target sizes through the RDM.

3.3.2. Low-Level Semantic Enhancement Module (LSEM)

In SAR images, aircraft usually appear near complex objects with strong scattering points, such as covered bridges and buildings, which cause great interference to aircraft detection. We hope to better utilize low-level semantic information to suppress this interference. Unfortunately, the existing CNN algorithms mainly extract high-level semantic features of objects. By contrast, the high-level semantic features of SAR images are far less abundant than those of optical images. In order to better learn the spatial location information and scattering characteristics of SAR aircraft targets, this paper draws on the idea of inception [50] and introduces the coordinate attention mechanism (CoordAtt) [47] to design a low-level semantic enhancement module (LSEM). The specific structure of the LSEM is shown in Figure 7. The Conv block is used as the fundamental component of design. First, the C3 feature map is used as the input to reduce the dimension through the Conv block with the convolution kernel of 1 on the four branches, and then feature extraction is performed in different ways for each branch. Finally, the output features of the four branches are stitched to achieve the effect of multi-feature fusion of low-level semantic information.

3.3.3. Localization Attention Module (LAM)

The fused feature map of the RDM and the LSEM already has enough information for aircraft target detection in complex environments. However, the feature information is redundant and easily confused due to surrounding objects. To this end, we design a localization attention module (LAM) to refine the fused feature map, hoping to achieve the effect that aircraft targets differ significantly from the surrounding features on different feature channels. As shown in Figure 8, the module first takes the fused output feature X as input through two branches. A branch first reduces the dimension of the input channel through a 1 × 1 Conv. It then refines the semantic features through a bottleneck block to perform a splicing operation with the other branch’s 1 × 1 Conv channel dimension-reduction feature. The bottleneck block is a residual block formed by stacking two Conv blocks. Then, the output features are enhanced with spatial localization information by CoordAtt. Finally, the channel dimension and the nonlinearity of the network are maintained by a Conv block. In this way, LAM refines the features, strengthens the target’s response on the feature map, and suppresses interference due to surrounding objects.

4. Experiments and Analysis

4.1. Dataset Description

The Gaofen-3 aircraft target dataset [51,52] in this paper has been collected by the Gaofen-3 satellite and consists of single-polarization SAR image with a resolution of 1 m in the C-band and includes multi-temporal phase maps of multiple airports. The dataset has a total of 2000 image slices, with image sizes ranging from 600 to 2048 pixels. It mainly includes seven types of civil aircraft, such as the Boeing 737, with a total of 6556 aircraft samples. Figure 9 shows from the histogram of the bounding box distribution that there is only one aircraft in many images, but at most there are 35 aircraft in a picture. Figure 10 shows a partial dataset slice, with the area marked by the green box the aircraft target. Furthermore, strong scattering points such as covered bridges and buildings are around the aircraft. By combining Figure 9, Figure 10 and Figure 11, it can be seen that the aircraft targets in this dataset have uneven image quality, significant differences in the sizes of aircraft targets, dense target arrangement, and complex surrounding ground objects. This paper randomly divides the dataset into training, validation, and test sets at a ratio of 6:2:2 for experiments.

4.2. Experimental Parameter Settings

The training image size is 640 × 640 pixels, and simple data augmentations such as panning, scaling, cropping, and flipping are done on the samples before training to enhance the model’s generalization. Four groups of anchor boxes are preset as: (16, 14), (27, 25), (52, 50), (90, 83). Due to the corresponding changes to the backbone network in this paper, the network weights in all ablation experiments are initialized randomly. The optimizer for model training is SGD, with the momentum factor size of 0.937, the initial learning rate of 0.01, and the weight decay of 0.0005. We perform a learning rate warm-up in the first three epochs to maintain a better gradient. Each experiment is performed for 300 epochs, and the model with the best evaluation on the validation set is reserved for the final training result. All experiments are carried out under the Pytorch1.9 deep learning framework on an Ubuntu 16.04 system with two Tesla M60 (16 GB) GPUs.

4.3. Evaluation Metrics

In order to better evaluate the performance of each algorithm on aircraft target detection, we adopt some common evaluation metrics for object-detection tasks [9,53]: precision (P), recall (R), F1 score, and average precision (AP) to measure the detection performance of the algorithm; and model parameters (Params), floating-point operations per second (FLOPs), and frames per second (FPS) to measure model complexity and inference speed. In the comparison algorithm, we also draw the precision–recall (PR) curve to show the detection performance of each algorithm.
IoU = 0.5 is used as the threshold for dividing positive and negative samples in the experimental evaluation. Precision (P) and recall (R) are, respectively, defined in Formulas (3) and (4), where TP, FP, TN, and FN denote true case, false positive case, true negative case, and false negative case, respectively.
P = T P T P + F P
R = T P T P + F N
F1 score is based on the harmonic mean of P and R. The larger the value of the score, the better the model’s performance. It is defined in Formula (5):
F 1 = 2 × P × R P + R
Average precision (AP) is the area enclosed by the PR curve and the coordinate axis. It comprehensively considers the effects of precision (P) and recall (R) to reflect the quality of the model. The larger the AP value, the better the model performance. Its definition is Formula (6):
AP = 0 1 p ( r ) dr
Model parameters (Params) measure the space complexity of the model and its corresponding memory resource occupation. For example, assuming that the current convolution layer uses K, M, and N to represent the size of the convolution kernel, the number of input channels, and the number of output channels, respectively, then this convolutional layer parameter quantity is defined in Formula (7):
Params = K 2 × M × N
Floating-point operations per second (FLOPs) measures the time complexity of the model, which is a reference indicator of the calculation time. Assuming that the current convolutional layer is represented by K, M, N, H, and W for the convolutional kernel size, input channels, output channels, and the height and width of the output feature map, respectively, the FLOPs of this convolutional layer is defined in Formula (8):
FLOP s = 2 K 2 × M × N × H × W
Frames per second (FPS) measures the overall detection speed of the algorithm and is defined in Formula (9), which gives the average detection time of an image. A larger value represents faster detection.
FPS = 1 t

4.4. Experimental Results and Analysis

4.4.1. Selection of Preset Anchor Boxes

Since the method in this paper does not use FPN as the neck and only detects at one resolution of 16-times downsampling, if the original scale of the anchor box preset is used, anchor box recall on the dataset is low, and there are many missed detections. Therefore, it is necessary to redesign a set of anchor boxes based on the dataset’s labeling information to optimize the model’s performance. Figure 11 shows the aspect ratio of the aircraft bounding boxes relative to the images; it can be seen that the aspect ratio of the aircraft target is close to 1:1, and most of the aircraft are small relative to the image. From a relative scale perspective, a target is usually defined as small when the ratio of the target’s bounding box area to the image area is less than 0.58% [49]. Data analysis finds that small targets account for 48.8% of the SAR aircraft dataset. For this reason, we determine the bounding box size of the training samples by the K-means clustering method and finally find four groups of anchor-box presets suitable for aircraft samples: (16, 14), (27, 25), (52, 50), (90, 83).

4.4.2. Ablation Experiments and Analysis

This section is a series of ablation experiments on the SEAN network architecture proposed in this paper to evaluate the effectiveness of each module based on evaluation metrics. The experiments are divided into three parts: choice of model size and depth, module effectiveness, and comparison between different backbone networks and necks on YOLOv5s.
(1) Selection of Model Size and Depth: This paper conducts experiments using each of the five model sizes provided by YOLOv5 version 6.0. As shown in Table 2, YOLOv5s has the best trade-off between accuracy and speed, and a short training time (T-time). Secondly, this paper explores the depth of the model, that is, which stage output features are more suitable for SAR aircraft target detection. As shown in Table 3, detection using only the C4 feature has a sufficiently high AP of 95.4%, which is 14.3% and 1.1% higher than that of the C3 and C5 features, indicating that the C4 feature has a good trade-off between resolution and high-level semantic information. On the other hand, the FPS is low when only the C3 feature is used for detection, although the FLOPs are minimal. This is due to the lack of high-level semantic information in the C3 feature map for the binary classification of ground objects and aircraft targets, which results in many false alarms in the prediction. This leads to short network inference time when using only the C3 features for detection, but too many false alarms are generated and too much time is spent on NMS post-processing of the anchor boxes. Based on the above experimental analysis, we select the C4 feature map of YOLOv5s plus the head as the basis for subsequent experiments.
(2) Effectiveness of Each Module: In order to clearly express the contribution of each module of SEAN, we conduct experiments based on the C4 plus the head. The results are shown in Table 4. Firstly, the structural re-parameterization technique is introduced into the backbone, i.e., RepConv is used instead of simple convolutional downsampling. It can be seen that the model does not affect the detection FPS, and the AP is improved by 0.6%. Then, for the neck design, we use RDM to fuse different receptive-field scale features to achieve 0.3% AP improvement, and LSEM to enhance the learning of low-level semantic information to achieve 0.4% AP improvement. Since the semantic information of directly splicing the output features of both RDM and LSEM is more confusing and redundant, LAM is added in this paper to refine the semantic information, which improves the AP by 1% by enhancing aircraft positioning information. Moreover, the AP of the loss function of SIoU [54] using the frontier later decreases by 0.4%. Figure 12 shows (a) the actual detection map of the method in this paper; (b) in the absence of LAM, the aircraft target is easily confused with the surrounding features on these grayscale channel maps; and (c) with LAM, the corresponding performance of the aircraft target on the grayscale channel maps is significantly different from that of the surrounding features. It can be seen that LAM can refine the semantic features and effectively suppress the interference of complex features on aircraft target detection. The attention mechanism can compensate for the problem of strong locality and insufficient globality of CNN by obtaining global context information. Table 5 shows the effects of different attention mechanisms on the performance of LAM. We added different attention modules in the same position based on the LAM without attention mechanism to verify the effect. The results show that almost all of these attention mechanisms improve the performance of the network; specifically, the overall performance of our adopted CoordAtt is the best. In general, the algorithm in this paper significantly reduces the parameters and FLOPs compared to YOLOv5s and improves the AP by 1.3% and 8.7 FPS on the test set. Figure 13 shows the change to AP on the validation set with increasing epochs during model training. Since SEAN only uses one scale of detection head, the model fitting speed is slower than that of YOLOv5s, but the final AP value is better.
To better understand the points of interest of the SEAN model under complex ground conditions, this paper uses the Grad-CAM [56] technique for visualization. This is a back-propagation gradient of the information from the predicted anchor boxes, and the learning of the model is represented by the response of the parameters of the last convolution layer using a heat map. In Figure 14a–c represent three cases of difficulty detecting due to complex ground objects, dense aircraft arrangement, and small aircraft targets. The actual detection effect of the model is in the upper half, and the corresponding Grad-CAM feature visualization is in the lower half. The figure shows that the model can accurately detect the aircraft targets in all these scenarios. Furthermore, the heat map shows that the areas with aircraft targets are red. In contrast, other objects such as corridors, buildings, and runways are presented in cool colors, which proves that our algorithm has strong anti-interference ability and high detection accuracy in complex scenes.
(3) Different Neck and Backbone: To further demonstrate the performance advantage of the proposed method, we substitute the backbone and neck of YOLOv5s for comparison with SEAN. For the backbone, we use ConvNeXt [57], and we train its tiny version with the same number of channels per module control as with YOLOv5s. Furthermore, the neck part is changed to FPN [13] and BiFPN [40]. Table 6 shows that our proposed algorithm has better accuracy and speed than YOLOv5s using different backbones and necks.

4.4.3. Comparative Experiments and Analysis

To further demonstrate the effectiveness of SEAN, we compare it to seven typical detection algorithms based on MMDetection [58]. As shown in Table 7, the long training time of relatively large backbone networks such as ResNet, especially two-stage algorithms, makes them prone to over-fitting on small-scale SAR data. Therefore, we use a pre-trained model for initialization in the backbone of these networks. In terms of detection speed, the table shows that models using ResNet50 as the backbone and FPN as the neck, such as Cascade R-CNN, have massive computational effort, resulting in significantly lower detection speeds. Moreover, detection speed is improved when YOLOF adopts the dilated encoder (D-en) as the neck. While YOLOv3, YOLOX-s, and SEAN (our method) all use a lightweight backbone, SEAN uses a simple neck to significantly improve detection speed. Regarding detection accuracy, YOLOv3, YOLOX-s, and SEAN (our method), which do not use ResNet50 as the backbone, are generally better than the other methods. As shown in Figure 15, the area of the PR curve of the SEAN algorithm is significantly larger than that of the other comparison algorithms, which means that the method in this paper has a significant detection accuracy advantage in actual detection. According to the above experiments, the proposed SEAN method has better average accuracy of 97.7%, a better F1 score of 94.9%, and faster detection speed of 83.3 FPS with fewer parameters and FLOPs than the typical algorithms on the SAR aircraft dataset.
In order to visualize the effectiveness of SEAN, we select some challenging detection scenarios from the test set to compare the algorithms. Figure 16 shows a small target scene, and SEAN accurately detects the small aircraft target. Figure 17 and Figure 18 are two scenes with complex environments; SEAN has significantly fewer missed targets and false alarms compared to other models and can effectively avoid the influence of strong scattering points due to complex objects. Moreover, among the aircraft targets detected by SEAN, the bounding boxes completely wrap whole aircraft, and there are no situations in which only local components are detected.

5. Conclusions

In this paper, we propose a simple and efficient attention network (SEAN) for aircraft detection in SAR images, which avoids the previous deep backbone and complex, laterally connected FPN neck and improves the detection accuracy and speed while significantly reducing the parameters and the FLOPs in the network. Through experiments, SEAN achieves 97.7% AP and 83.3 FPS speed on the Gaofen-3 aircraft target dataset. The results show that this algorithm has apparent advantages in detection accuracy and speed for SAR aircraft targets in complex backgrounds compared to other typical target algorithms. It shows that the trained model with a large amount of SAR aircraft labeled data has a high enough detection accuracy. However, due to the difficulty of manual labeling of SAR aircraft targets and the small amount of SAR data, the detection of SAR aircraft with small samples and the use of self-supervised learning methods in the field of SAR is very promising. Furthermore, we also find that the C4 feature of the backbone is more suitable for aircraft detection, and it can also be used to conduct lightweight network research on SAR in the future.

Author Contributions

Methodology, P.H. and D.L.; Project administration, P.H. and Z.C.; Software, P.H. and D.L.; Validation, P.H., D.L., B.H. and Z.C.; Writing, P.H., D.L. and B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Central University Basic Scientific Research Project of China (no. 3122020043).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, J.S.; Pottier, E. Polarimetric Radar Imaging: From Basics to Applications; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
  2. Guo, Q.; Wang, H.; Xu, F. Research progress on aircraft detection and recognition in SAR imagery. J. Radars 2020, 9, 497–513. [Google Scholar]
  3. Finn, H.M. Adaptive detection mode with threshold control as a function of spatially sampled clutter level estimates. RCA Rev. 1968, 29, 414–463. [Google Scholar]
  4. Gao, G.; Liu, L.; Zhao, L.; Shi, G.; Kuang, G. An Adaptive and Fast CFAR Algorithm Based on Automatic Censoring for Target Detection in High-Resolution SAR Images. IEEE Trans. Geosci. Remote Sens. 2009, 47, 1685–1697. [Google Scholar] [CrossRef]
  5. Zhang, H.; Hao, M.; Zhang, C.; Su, X. Aircraft target detection algorithm based on high resolution spaceborne SAR imagery. In Proceedings of the MIPPR 2017: Remote Sensing Image Processing, Geographic Information Systems, and Other Applications, Xiangyang, China, 28–29 October 2017; SPIE: Bellingham, WA, USA, 2018; Volume 10611, pp. 193–198. [Google Scholar]
  6. Han, P.; Wu, R.; Wang, Z.; Wang, Y. SAR automatic target recognition based on KPCA criterion. J. Electron. Inf. Technol. 2003, 25, 1297–1301. [Google Scholar]
  7. Ma, F.; Zhang, F.; Yin, Q.; Xiang, D.; Zhou, Y. Fast SAR image segmentation with deep task-specific superpixel sampling and soft graph convolution. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5214116. [Google Scholar] [CrossRef]
  8. Huang, Z.; Dumitru, C.O.; Pan, Z.; Lei, B.; Datcu, M. Classification of large-scale high-resolution SAR images with deep transfer learning. IEEE Geosci. Remote Sens. Lett. 2020, 18, 107–111. [Google Scholar] [CrossRef]
  9. Kang, Y.; Wang, Z.; Fu, J.; Sun, X.; Fu, K. SFR-Net: Scattering feature relation network for aircraft detection in complex SAR images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5218317. [Google Scholar] [CrossRef]
  10. He, C.; Tu, M.; Xiong, D.; Tu, F.; Liao, M. A component-based multi-layer parallel network for airplane detection in SAR imagery. Remote Sens. 2018, 10, 1016. [Google Scholar] [CrossRef]
  11. Guo, Q.; Wang, H.; Xu, F. Scattering enhanced attention pyramid network for aircraft detection in SAR images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7570–7587. [Google Scholar] [CrossRef]
  12. Zhao, Y.; Zhao, L.; Li, C.; Kuang, G. Pyramid attention dilated network for aircraft detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2020, 18, 662–666. [Google Scholar] [CrossRef]
  13. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  14. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 Octomber–2 November 2019; pp. 9627–9636. [Google Scholar]
  15. Zhao, Y.; Zhao, L.; Kuang, G. Fast detection aircrafts in complex large scene SAR images. Chin. J. Radio Sci. 2020, 35, 594–602. [Google Scholar]
  16. Zhang, L.; Li, C.; Zhao, L.; Xiong, B.; Quan, S.; Kuang, G. A cascaded three-look network for aircraft detection in SAR images. Remote Sens. Lett. 2019, 11, 57–65. [Google Scholar] [CrossRef]
  17. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  18. Zhang, P.; Xu, H.; Tian, T.; Gao, P.; Tian, J. SFRE-Net: Scattering Feature Relation Enhancement Network for Aircraft Detection in SAR Images. Remote Sens. 2022, 14, 2076. [Google Scholar] [CrossRef]
  19. Zhao, Y.; Zhao, L.-J.; Kuang, G.-Y. Attention feature fusion network for rapid aircraft detection in SAR images. ACTA Electonica Sin. 2021, 49, 1665. [Google Scholar]
  20. Luo, R.; Xing, J.; Chen, L.; Pan, Z.; Cai, X.; Li, Z.; Wang, J.; Ford, A. Glassboxing Deep Learning to Enhance Aircraft Detection from SAR Imagery. Remote Sens. 2021, 13, 3650. [Google Scholar] [CrossRef]
  21. Wang, J.; Xiao, H.; Chen, L.; Xing, J.; Pan, Z.; Luo, R.; Cai, X. Integrating Weighted Feature Fusion and the Spatial Attention Module with Convolutional Neural Networks for Automatic Aircraft Detection from SAR Images. Remote Sens. 2021, 13, 910. [Google Scholar] [CrossRef]
  22. Luo, R.; Chen, L.; Xing, J.; Yuan, Z.; Tan, S.; Cai, X.; Wang, J. A Fast Aircraft Detection Method for SAR Images Based on Efficient Bidirectional Path Aggregated Attention Network. Remote Sens. 2021, 13, 2940. [Google Scholar] [CrossRef]
  23. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  24. Li, J.; Qu, C.; Shao, J. Ship detection in SAR images based on an improved faster R-CNN. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–14 November 2017; IEEE: Washington, DC, USA, 2017; pp. 1–6. [Google Scholar]
  25. Liao, L.; Du, L.; Guo, Y. Semi-supervised SAR target detection based on an improved faster R-CNN. Remote Sens. 2021, 14, 143. [Google Scholar] [CrossRef]
  26. Niu, S.; Qiu, X.; Lei, B.; Ding, C.; Fu, K. Parameter extraction based on deep neural network for SAR target simulation. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4901–4914. [Google Scholar] [CrossRef]
  27. Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; Sun, J. RepVGG: Making VGG-Style ConvNets Great Again. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 13728–13737. [Google Scholar]
  28. Cai, Z.; Vasconcelos, N. Cascade R-CNN: Delving Into High Quality Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6154–6162. [Google Scholar]
  29. Ren, S.; He, K.; Girshick, R.B.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 1137–1149. [Google Scholar] [CrossRef]
  30. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  31. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  32. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  33. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  34. Li, Y.; Mao, H.; Girshick, R.B.; He, K. Exploring Plain Vision Transformer Backbones for Object Detection. arXiv 2022, arXiv:2203.16527. [Google Scholar]
  35. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  36. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  37. Jocher, G.; Stoken, A.; Chaurasia, A.; Borovec, J.; Kwon, Y.; Michael, K.; Liu, C.; Fang, J.; Abhiram, V.; Skalski, S.P. Ultralytics/yolov5: V6.0—YOLOv5n ‘Nano’ models, Roboflow integration, TensorFlow export, OpenCV DNN support. Zenodo Tech. Rep. 2021. Available online: https://zenodo.org/record/5563715#.YyglfLRByUk (accessed on 10 July 2022).
  38. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  39. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  40. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10778–10787. [Google Scholar]
  41. Ghiasi, G.; Lin, T.Y.; Pang, R.; Le, Q.V. NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 7029–7038. [Google Scholar]
  42. Chen, Q.; Wang, Y.; Yang, T.; Zhang, X.; Cheng, J.; Sun, J. You Only Look One-level Feature. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13034–13043. [Google Scholar]
  43. Gao, Z.; Wang, L.; Han, B.; Guo, S. AdaMixer: A Fast-Converging Query-Based Object Detector. arXiv 2022, arXiv:2203.16507. [Google Scholar]
  44. Guo, M.H.; Xu, T.X.; Liu, J.J.; Liu, Z.N.; Jiang, P.T.; Mu, T.J.; Zhang, S.H.; Martin, R.R.; Cheng, M.M.; Hu, S.M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  45. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  46. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018. [Google Scholar]
  47. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar]
  48. Chen, X.; Gong, Z. YOLOv5-Lite: Lighter, Faster and Easier to Deploy. 2021. Available online: https://zenodo.org/record/5241425/export/hx#.YyglU7RByUk (accessed on 10 July 2022).
  49. Chen, C.; Liu, M.Y.; Tuzel, O.; Xiao, J. R-CNN for Small Object Detection. In Proceedings of the ACCV, Taipei, Taiwan, 20–24 November 2016. [Google Scholar]
  50. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  51. 2021 Gaofen Challenge on Automated High-Resolution Earth Observation Image Interpretation. Available online: http://gaofen-challenge.com (accessed on 10 July 2022).
  52. Sun, X.; Lv, Y.; Wang, Z.; Fu, K. SCAN: Scattering Characteristics Analysis Network for Few-Shot Aircraft Classification in High-Resolution SAR Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5226517. [Google Scholar] [CrossRef]
  53. Everingham, M.; Eslami, S.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  54. Gevorgyan, Z. SIoU Loss: More Powerful Learning for Bounding Box Regression. arXiv 2022, arXiv:2205.12740. [Google Scholar]
  55. Misra, D.; Nalamada, T.; Arasanipalai, A.U.; Hou, Q. Rotate to attend: Convolutional triplet attention module. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 3139–3148. [Google Scholar]
  56. Selvaraju, R.R.; Das, A.; Vedantam, R.; Cogswell, M.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. Int. J. Comput. Vis. 2019, 128, 336–359. [Google Scholar] [CrossRef] [Green Version]
  57. Liu, Z.; Mao, H.; Wu, C.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 11976–11986. [Google Scholar]
  58. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  59. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. TOOD: Task-Aligned One-Stage Object Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; IEEE Computer Society: Washington, DC, USA, 2021; pp. 3490–3499. [Google Scholar]
  60. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
Figure 1. The architecture of YOLOv5 version 6.0.
Figure 1. The architecture of YOLOv5 version 6.0.
Remotesensing 14 04669 g001
Figure 2. Coordinate Attention (CoordAtt).
Figure 2. Coordinate Attention (CoordAtt).
Remotesensing 14 04669 g002
Figure 3. The overall architecture of the proposed SEAN. The network includes the backbone, the neck, and the head.
Figure 3. The overall architecture of the proposed SEAN. The network includes the backbone, the neck, and the head.
Remotesensing 14 04669 g003
Figure 4. Structural re-parameterization Conv block. K and S represent the size and stride of the convolution kernel. (a) Structural change process. (b) Re-parameterization process.
Figure 4. Structural re-parameterization Conv block. K and S represent the size and stride of the convolution kernel. (a) Structural change process. (b) Re-parameterization process.
Remotesensing 14 04669 g004
Figure 5. The structure of the RDM. K and D represent the size and dilation of the convolution kernel.
Figure 5. The structure of the RDM. K and D represent the size and dilation of the convolution kernel.
Remotesensing 14 04669 g005
Figure 6. Size of aircraft covered by feature information. (a) The receptive field of the C4 feature. (b) The receptive field after dilated convolution. (c) The receptive field after the RDM.
Figure 6. Size of aircraft covered by feature information. (a) The receptive field of the C4 feature. (b) The receptive field after dilated convolution. (c) The receptive field after the RDM.
Remotesensing 14 04669 g006
Figure 7. The structure of the LSEM. K is the size of the convolution kernel.
Figure 7. The structure of the LSEM. K is the size of the convolution kernel.
Remotesensing 14 04669 g007
Figure 8. The structure of the LAM. K is the size of the convolution kernel.
Figure 8. The structure of the LAM. K is the size of the convolution kernel.
Remotesensing 14 04669 g008
Figure 9. Distribution of the number of bounding boxes per image.
Figure 9. Distribution of the number of bounding boxes per image.
Remotesensing 14 04669 g009
Figure 10. Part of the training samples. The green boxes represent the ground truth of the aircraft samples.
Figure 10. Part of the training samples. The green boxes represent the ground truth of the aircraft samples.
Remotesensing 14 04669 g010
Figure 11. Relative size distribution of the bounding boxes. Dark blue dots indicate the presence of multiple aircraft targets of the same size.
Figure 11. Relative size distribution of the bounding boxes. Dark blue dots indicate the presence of multiple aircraft targets of the same size.
Remotesensing 14 04669 g011
Figure 12. Channel feature visualization. (a) The actual detection map of the method in this paper. (b) Feature channel map without LAM. (c) Feature channel map with LAM.
Figure 12. Channel feature visualization. (a) The actual detection map of the method in this paper. (b) Feature channel map without LAM. (c) Feature channel map with LAM.
Remotesensing 14 04669 g012
Figure 13. Training curves on the validation set. We evaluate the AP of each epoch for model training, and the results show that SEAN is better than YOLOv5s.
Figure 13. Training curves on the validation set. We evaluate the AP of each epoch for model training, and the results show that SEAN is better than YOLOv5s.
Remotesensing 14 04669 g013
Figure 14. Grad-CAM visualization of three scenes (ac); the upper part is the prediction result of the SEAN model, and the bottom part is the visualized heat map.
Figure 14. Grad-CAM visualization of three scenes (ac); the upper part is the prediction result of the SEAN model, and the bottom part is the visualized heat map.
Remotesensing 14 04669 g014
Figure 15. The PR curves of various methods.
Figure 15. The PR curves of various methods.
Remotesensing 14 04669 g015
Figure 16. Scene 1. Small aircraft target detection.
Figure 16. Scene 1. Small aircraft target detection.
Remotesensing 14 04669 g016
Figure 17. Scene 2. Aircraft target detection in complex environment.
Figure 17. Scene 2. Aircraft target detection in complex environment.
Remotesensing 14 04669 g017
Figure 18. Scene 3. Aircraft target detection in complex environment.
Figure 18. Scene 3. Aircraft target detection in complex environment.
Remotesensing 14 04669 g018
Table 1. Network configuration of the proposed SEAN.
Table 1. Network configuration of the proposed SEAN.
SequenceFromnModuleArguments
1−11Conv[3, 32, 6, 2, 2]
2−11RepConv[32, 64, 3, 2]
3−11CSPC3[64, 64]
Backbone4−11RepConv[64, 128]
5−12CSPC3[128, 128]
6−11RepConv[128, 256, 3, 2]
7−13CSPC3[256, 256]
8−11RDM[256, 256]
951LSEM[128, 128]
Neck10−11RepConv[128, 256, 3, 2]
11[−1, 7, 8]1Concat[1]
12−11LAM[756, 512]
Head13−11YOLOHead[512, 1]
Table 2. Selection of model size.
Table 2. Selection of model size.
MethodP (%)R (%)F1 (%)AP (%)Params (M)FLOPs (G)T-Time (h)FPS
YOLOv5n93.891.692.795.61.774.20.84123.3
YOLOv5s93.792.493.096.47.0116.01.1574.6
YOLOv5m93.594.293.896.520.8748.02.6036.1
YOLOv5l94.193.894.096.446.11107.84.6722.7
YOLOv5x93.495.394.496.686.12204.29.1913.1
Table 3. Selection of backbone depth.
Table 3. Selection of backbone depth.
MethodP (%)R (%)F1 (%)AP (%)Params (M)FLOPs (G)FPS
C3+Head81.872.476.881.10.235.158.8
C4+Head93.392.192.795.41.158.0107.5
C5+Head94.487.390.794.33.529.9104.2
Table 4. Ablation experiments of each SEAN module.
Table 4. Ablation experiments of each SEAN module.
MethodP (%)R (%)F1 (%)AP (%)Params (M)FLOPs (G)FPS
YOLOv5s93.792.493.096.47.0116.074.6
C4+Head93.392.192.795.41.158.0107.5
+RepConv91.694.192.896.0 (+0.6)1.158.0107.5
+RDM94.792.793.796.3 (+0.3)1.168.1103.2
+LSEM93.793.593.696.7 (+0.4)1.519.692.6
+LAM (ours)95.694.294.997.7 (+1.0)2.8413.883.3
+SIoU [54]94.295.594.997.3 (−0.4)2.8413.882.6
Table 5. The impact of different attention mechanisms on LAM.
Table 5. The impact of different attention mechanisms on LAM.
MethodP (%)R (%)F1 (%)AP (%)
LAM (Basic)94.694.594.597.2
LAM (SE [51])94.994.494.697.3
LAM (CBAM [54])93.794.694.197.5
LAM (TripletAtt [55])93.995.194.597.2
LAM (CoordAtt [47])95.694.294.997.7
Table 6. Comparison to YOLOv5 with different backbones and necks.
Table 6. Comparison to YOLOv5 with different backbones and necks.
MethodBackboneNeckF1 (%)AP (%)Params (M)FLOPs (G)FPS
SEAN (ours)Modified CSPDarknetOurs94.997.72.8413.883.3
YOLOv5sCSPDarknetPANet93.096.47.0116.074.6
CSPDarknetFPN92.796.06.8515.878.2
CSPDarknetBiFPN93.696.57.0716.174.2
ConvNeXt [57]PANet93.897.025.352.716.9
Table 7. Comparison to different algorithms.
Table 7. Comparison to different algorithms.
MethodBackboneNeckF1 (%)AP (%)Params (M)FLOPs (G)FPS
Faster R-CNNResNet50FPN88.986.141.12182.311.2
Cascade R-CNNResNet50FPN87.384.568.93237.67.3
FCOSResNet50FPN89.183.531.84153.715.3
YOLOFResNet50D-en83.879.442.0678.522.3
TOOD [59]ResNet50FPN91.188.831.79144.312.9
YOLOv3MobileNetv2 [60]FPN94.892.73.7413.544.2
YOLOX-sCSPDarknetPANet94.093.08.9426.341.2
SEAN (ours)Modified CSPDarknetOurs94.997.72.8413.883.3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Han, P.; Liao, D.; Han, B.; Cheng, Z. SEAN: A Simple and Efficient Attention Network for Aircraft Detection in SAR Images. Remote Sens. 2022, 14, 4669. https://doi.org/10.3390/rs14184669

AMA Style

Han P, Liao D, Han B, Cheng Z. SEAN: A Simple and Efficient Attention Network for Aircraft Detection in SAR Images. Remote Sensing. 2022; 14(18):4669. https://doi.org/10.3390/rs14184669

Chicago/Turabian Style

Han, Ping, Dayu Liao, Binbin Han, and Zheng Cheng. 2022. "SEAN: A Simple and Efficient Attention Network for Aircraft Detection in SAR Images" Remote Sensing 14, no. 18: 4669. https://doi.org/10.3390/rs14184669

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop