Next Article in Journal
Using Artificial Neural Networks to Assess Earthquake Vulnerability in Urban Blocks of Tehran
Next Article in Special Issue
Water Body Extraction from Sentinel-2 Imagery with Deep Convolutional Networks and Pixelwise Category Transplantation
Previous Article in Journal
Multi-Channel Spectral Band Adjustment Factors for Thermal Infrared Measurements of Geostationary Passive Imagers
Previous Article in Special Issue
Quantitative Short-Term Precipitation Model Using Multimodal Data Fusion Based on a Cross-Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Slicing-Aided Hyper Inference for Small Object Detection in High-Resolution Remote Sensing Images

1
School of Education Science and Technology, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
2
Internet of Things & Smart City Innovation Platform, Zhuhai Fudan Innovation Institute, Zhuhai 519031, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(5), 1249; https://doi.org/10.3390/rs15051249
Submission received: 27 December 2022 / Revised: 18 February 2023 / Accepted: 21 February 2023 / Published: 24 February 2023

Abstract

:
In the field of object detection, deep learning models have achieved great success in recent years. Despite these advances, detecting small objects remains difficult. Most objects in aerial images have features that are a challenge for traditional object detection techniques, including small size, high density, high variability, and varying orientation. Previous approaches have used slicing methods on high-resolution images or feature maps to improve performance. However, existing slicing methods inevitably lead to redundant computation. Therefore, in this article we present a novel adaptive slicing method named ASAHI (Adaptive Slicing Aided Hyper Inference), which can dramatically reduce redundant computation using an adaptive slicing size. Specifically, ASAHI focuses on the number of slices rather than the slicing size, that is, it adaptively adjusts the slicing size to control the number of slices according to the image resolution. Additionally, we replace the standard non-maximum suppression technique with Cluster-DIoU-NMS due to its improved accuracy and inference speed in the post-processing stage. In extensive experiments, ASAHI achieves competitive performance on the VisDrone and xView datasets. The results show that the mAP50 is increased by 0.9% and the computation time is reduced by 20–25% compared with state-of-the-art slicing methods on the TPH-YOLOV5 pretrained model. On the VisDrone2019-DET-val dataset, our mAP50 result is 56.4% higher, demonstrating the superiority of our approach.

Graphical Abstract

1. Introduction

In recent years, object detection has received much attention due to its widespread applications in various fields. When applied to images from nature, traditional object detectors such as RetinaNet [1], YOLOV5, and Fast-RCNN [2] have achieved impressive results. However, their results are not satisfactory for drone-captured and remote sensing images. For example, Faster R-CNN [3], a classical two-stage object detection algorithm, achieves 38.0 mAP on objects from the VisDrone2019 evaluation set with various sizes, but only 6.2 and 20.0 mAP, respectively, on objects with tiny and small sizes. This limitation is mainly caused by three factors: (1) low-resolution features in the deeper layer may not match the diverse size of small objects; (2) qualities such as high densities, large numbers of objects, variable scales, and small occupied areas cause significant challenges; and (3) detecting small objects is more difficult compared with large objects, as the Intersection over Union (IoU) metric can fail to account for situations such as overlapping objects, close proximity between objects, and diverse aerial perspectives.
To tackle these problems, specialized small object detection algorithms have been developed. With the appearance of 4K cameras and civilian drones and the fast development of deep learning and remote sensing technologies, the performance of small object detection using convolutional neural networks (CNNs) has substantially improved. Several methods, such as TPH-YOLOV5, TOOD [4], PP-YOLOE [5], UAVs, and SODNet [6] have obtained impressive results and are widely employed in tasks such as geographical surveys [7], person identification [8], flow monitoring [9], and traffic control. Most of these tasks require processing of high-resolution images. Hence, a number of researchers have attempted to increase the upsampling rate to maintain high-resolution features, for example, in algorithms such as TPH-YOLOV5 [10] and QueryDet [11]. Meanwhile, adding penalty terms in the post-processing stage can help to tackle problems associated with high densities. Several post-processing methods, such as Soft-NMS [12], WBF [13], and DIoU-NMS [14], have shown effectiveness in terms of improved performance. In addition, the slicing method [15] and the shifted window-based self-attention mechanism [16] have been proposed; these approaches slice images into smaller patches to amplify their local features. Both of these slicing methods have been widely used in small object detection algorithms to achieve better performance [10,11,15]. However, the redundant computations of the traditional slicing method remain, increasing computational costs and reducing detection speed. Thus, in this work our approach focuses on alleviating redundant problems to obtain better accuracy and faster detection speed when employing a small object detection algorithm on remote sensing images.
In this paper, we propose a novel slicing method, ASAHI (Adaptive Slicing-Aided Hyper Inference), which adaptively slices images into a fixed number of patches rather than using a fixed slicing size. By automatically controlling the number of slices, our method greatly reduces redundant computation. In the post-processing stage, we substitute non-maximum suppression (NMS) with Cluster-DIoU-NMS, which reduces the time consumption while maintaining the quality of the results. This approach was motivated by two key observations. First, in SAHI [15], we found that a fixed slice size leads to different degrees of redundant calculation for images with different resolutions. In particular, at the edge of the image, the redundant calculation rate exceeds the overlapping rate. By contrast, adaptive slicing size is preferable. Second, small object detection most contend with the difficulties posed by close proximity, overlap, and occlusion. NMS is a mainstream post-processing method that is widely used for object detection; however, it does not perform well for high-density objects due to its reliance on a single metric, namely, IoU. This limitation is evident in small object detection. Therefore, combining Cluster-NMS [17] and DIoU-NMS [14] is a good scheme.
We evaluated ASAHI on a challenging dataset, VisDrone2019 [18], which contains a large quantity of small objects in different resolutions. In addition, we tested it on the xView [19] dataset, which is annotated with bounding boxes and contains large amounts of aerial imagery and complex scenes from around the world. Large-scale experiments show that our method accelerates the inference and post-processing stages by 20–25%, while improving the detection performance by 2.1% compared with the baseline SAHI method [15]. Figure 1 shows the improved results obtained by ASAHI.
The main contributions of this paper are as follows:
  • We propose a novel slicing method, ASAHI, which can improve detection performance for images of various resolutions and targets at a range of scales by solving the redundancy problems in the previous SAHI method [15].
  • We combine ASAHI with the TPH-YOLOV5 framework [10] to strengthen its sensitivity to the lower features in deeper layers.
  • We design an adaptive slicing-aided fine-tuning scheme called ASAF for data augmentation, which is more suitable for ASAHI than other traditional fine-tuning strategies.
  • We integrate Cluster-DIoU-NMS, a faster Cluster-NMS method with DIoU penalty, in the post-processing stage to improve the accuracy and detection speed of inferences in high-density scenarios.

2. Related Work

Object Detection. The recent CNN-based object detectors can be divided into two main types, namely, single-stage detectors and two-stage detectors. Single-stage detectors, such as YOLOv3 [20], SSD [21], RetinaNet [1], EfficientDet [22], and YOLOV4 [23], directly predict the location of the targets; single-stage detection can be regarded as a regression problem. Two-stage detection tends to use the RoIAlign operation [24] to align an object’s features explicitly. Several typical frameworks, such as R-CNN [25], SPPNet [26], DetectoRS [27], and FPN [28], utilize a region proposal stage to generate candidate objects and selectively search those parts of an image that have a high probability of belonging to an object. Typically, single-stage detectors are less computationally expensive, while two-stage detectors achieve higher accuracy. However, the performance gap between these two approaches has narrowed recently. For example, YOLOV5 is a classical single-stage detector that uses the CSPDarknet53 architecture, with an SPP layer as the backbone and PANet as the neck. During inference, large numbers of bounding boxes are obtained and then aggregated by CIoU in the post-processing stage. This approach has attracted the attention of researchers because of its convenience and good performance.
Small Object Detection. Small object detection is a challenging task in computer vision due to the difficulties posed by low resolution, high density, variable scales, and small occupation areas. A number of researchers have modified the traditional object detection algorithms with self-attention mechanism such as CBAM [29], SPANet [30] and RTANet [31] to accommodate small object detection. An excellent work is TPH-YOLOV5 [10]; a state-of-the-art small object detection framework based on YOLOV5, it integrates an additional prediction head for small object detection, a CBAM [29] framework in the neck, and replaces the original CSP bottleneck blocks with transformer encoder blocks. This increases the detection accuracy for small objects while preserving the speed of inference. Hence, in this work we utilize TPH-YOLOV5 [10] as our backbone and propose ASAHI to achieve better results.
Postprocessing in Small Object Detection. Post-processing is a crucial part of small object detection. During the inference process, a large number of bounding boxes are generated; however, many of them represent detections of the same target, and a post-processing method must be used to suppress these false detections. In recent years, the DETR method [32] has replaced the traditional post-processing process; however, the development of this technology is not yet complete. In this paper, we maintain the use of post-processing. NMS is one of the most popular post-processing methods. Several well-known frameworks, such as TPH-YOLOV5 [10] and SAHI [15], use NMS in the post-processing stage. Although, NMS is competitive, it has a number of shortcomings due to IoU being the only influence factor in NMS. Thus, adding penalty terms is an efficient way to improve performance or reduce time consumption. For example, GIoU, DIoU, and CIoU add penalties based on the weight of non-overlapping areas, the center point distance between bounding boxes, and the aspect ratio, respectively, to improve detection performance. In addition, several methods have been combined with NMS to achieve better performance in the post-processing stage, such as IBS [33], a state-of-the-art two-stage algorithm with excellent performance, Soft-NMS [12], a soft improvement based on NMS with Gaussian algorithm, and Cascade R-CNN + NWD [34], a classical improved algorithm based on the normalized Wasserstein distance. Our method, ASAHI, focuses on better performance and faster inference. Inspired by LD [35], we employed a combination of the Cluster-NMS and DIoU-NMS [14] algorithms to meet our goals.

3. The Proposed Method

In this paper, with the aim of enhancing small object detection accuracy, we propose a novel slicing method called ASAHI, an overview of which is shown in Figure 2. The overall process can be simplified as follows: two strategies, full inference (FI) and ASAHI, are applied to the original images simultaneously. In FI, the full images are sent to the inference stage. By contrast, ASAHI adaptively changes the slicing size to control the number of slices according to the image resolution. In the inference stage, the full image and sliced image are processed by TPH-YOLOV5 [10] to obtain the corresponding bounding boxes. Next, the bounding boxes are resized by the aspect ratio and merged using Cluster-DIoU-NMS to remove the repeatedly identified pre-selected boxes in the post-processing stage. In the remainder of this section, we introduce ASAHI, Cluster-DIoU-NMS, and Adaptive Slicing Aided Fine-tuning (ASAF) in detail.

3.1. Overview of TPH-YOLOV5

TPH-YOLOV5 [10] provides two pre-trained models, yolov5l-xs-1 and yolov5l-xs-2, which were trained at different sizes after resizing. TPH-YOLOV5 [10] uses the architecture of CSPDarknet53, with three transformer encoder blocks as the backbone, PANet with CBAM [29] as the neck, and four transformer prediction heads. Considering its excellent performance, ranking fifth in the VisDrone 2021 competition, we selected it as our backbone. Thanks to the code provided by the authors of the study [10], we used a slight modification in the neck, as shown in Figure 3, in which the number of CBAM blocks [29] is reduced by one.

3.2. Adaptive Slicing-Aided Hyper Inference (ASAHI)

By investigating a range of slicing methods, we found that fixed slice sizes inevitably lead to redundant computation in images with different resolutions. When a fixed slice size cannot be used exactly, part of the information from the previous image is included to ensure that the slice is complete; this strategy causes redundant computation, and can pose a problem for merging during post-processing. Our ASAHI approach, as an improved version of the SAHI slicing algorithm, greatly reduces this redundant computation by adaptively changing the slicing size.
During the inference step, as detailed in Figure 4, the original image is sliced into 6 or 12 overlapping patches according to the image size. Then, each patch is resized, preserving the aspect ratio. After that, the patches are transferred to the merge stage. However, as this form of slicing results in clear defects in the detection of large objects, we use FI to detect the larger objects. Finally, Cluster-DIoU-NMS is used to merge the FI results in order to return them to their original size.
The redundancy in the inference process is shown in Figure 5, in which W and H are the length and width of the input image, respectively, and l represents the overlapping ratio. In our experiments, we noticed that the performance, as measured by the mAP, decreases if the slice sizes are different in length and width. Hence, our method maintains the same slice size in terms of length and width.
In ASAHI, we set a differentiation threshold as LS: if the length or width of the image is larger than this threshold, then the image is cut into 12 patches; otherwise, it is cut into 6 patches. LS can be calculated according to Equation (1) below. We define the restricted slice size, denoted as restrict_size, to guarantee the slice size fluctuation within restrict_size; here we set it to 512.
L S = restrict _ size · ( 4 3 l ) + 1 .
After setting the value of LS, we make sure that our image will be divided into exactly 6 and 12 patches by calculating two extreme cases and taking the maximum value:
p = max ( W / ( 3 2 l ) + 1 , H / ( 2 l ) + 1 ) max ( W , H ) LS max ( W / ( 4 3 l ) + 1 , H / ( 3 2 l ) + 1 ) max ( W , H ) LS .
Then, we let a be the number of slices on the horizontal axis and b the number of slices on the vertical axis. They can be calculated as follows:
a = W p l p ( 1 l ) ,
b = H p l p ( 1 l ) .
We can then calculate the redundant areas in Figure 5 from the three known values of W, H, l, and p in Equation (2), as follows:
S area = W H
Redundancy _ x = p a p l ( a 1 ) W ,
Redundancy _ y = p b p l ( b 1 ) H ,
S redundancy = Redundancy _ x · H + Redundancy _ y · W Redundancy _ x · Redundancy _ y .
After calculating the redundancy of the ASAHI method, we can derive the reduction in redundant computational area compared to the SAHI [15] method as follows:
S redundancy ASAHI = W ( p ( 2 l ) H ) W ( 3 2 l ) H ( 2 l ) & max ( W , H ) LS H ( p ( 3 2 l ) W ) W ( 3 2 l ) < H ( 2 l ) & max ( W , H ) LS W ( p ( 3 2 l ) H ) W ( 4 3 l ) H ( 3 2 l ) & max ( W , H ) > LS H ( p ( 4 3 l ) W ) W ( 4 3 l ) < H ( 3 2 l ) & max ( W , H ) > LS
S Reduction Time = 1 ( S redundancy ASAHI + S area ) / ( S redundancy + S area )
The redundant computation generated by SAHI [15] is decreased through our method. The proportions of the reduced redundant computation for images of various resolutions are shown in Table 1. As a result of this reduction, ASAHI is able to achieve improved processing speed.

3.3. Cluster-DIoU-NMS

Post-processing is a crucial part of small object detection. Inspired by LD [35], we employed Cluster-DIoU-NMS in the post-processing stage due to the outstanding performance of Cluster-NMS [17] and DIoU-NMS [14]. Briefly, Cluster-DIoU-NMS is a combination of DIoU-NMS [14] and Cluster-NMS [17]. We used the structure of Cluster-NMS in order to speed up the detection speed, and replaced the IOU metric with DIoU to improve performance in overlapping scenarios. In other words, the oversuppression issue is addressed by Cluster-NMS and the DIoU parameters are employed to strengthen the accuracy.
DIoU-NMS: In the traditional NMS algorithm, IoU is the only factor considered. However, in practical applications it is difficult to distinguish objects by IoU alone when the distance between two objects is small. Therefore, the overlapping area of the two bounding boxes and the distance of the center point should be considered simultaneously in post-processing; despite this, traditional IoU only takes the overlapping area into account, ignoring the centroid distance. To address this issue, DIoU-NMS [14] incorporates the distance between the center points of the two boxes as a penalty term. If the IoU between two boxes is relatively large and the distance between the centers is relatively small, the boxes are deemed to be separate objects and are not filtered out. In DIoU, c 2 denotes the diagonal distance of the smallest wrapped box overlaying the two bounding boxes, while ρ 2 ( x , x g t ) represents the Euclidean distance between the centroids of the two boxes. DIoU is defined as
DIoU = IoU ρ 2 ( x , x g t ) c 2 .
Cluster-NMS: In traditional NMS, as shown in the left of Figure 6, the matrix X consists of i rows and j columns, which are sorted by score and upper triangulation, and each element X i j represents the computed IoU of the ith bbox and the jth bbox. Then, the maximum value is taken for each column and binarized to obtain the resulting sequence b . However, in the blue part of Figure 6, 0.6 is calculated from the first and second bbox; the IoU exceeds the threshold, and should be suppressed. The first bbox and the second bbox should be considered as one box. Therefore, the bbox in the fourth column of the second row should be excluded, which is computed by the second and fourth bbox. This means that 0.8 should not appear. In contrast, Cluster-NMS employs a simple and effective alteration. Before binarization, X C l u s t e r multiplies b to the left by X , which can solve the over-inhibition problem, as shown in the right of Figure 6.

3.4. Adaptive Slicing-Aided Fine Tuning

In ASAF, the widely used TPH-YOLOv5 object detection framework [10] provides pre-trained weights on datasets such as VisDrone. This allows the model to be fine-tuned using smaller datasets with shorter training times, rather than training from scratch with large datasets. These pre-trained models provide successful detection performance for similar inputs. In ASAF, as shown in Figure 7, I 1 , I 2 , , I j are sliced into overlapping patches P 1 , P 2 , , P j by ASAHI to form the sliced dataset. Together with the original pre-trained dataset, we set up our complete fine-tuned dataset. Considering the large number of images after slicing, which imposes a large burden on the training process, we have not used other data augmentation techniques such as rotation, geometric distortions, or photometric distortions.

4. Experiments

4.1. Datasets and Settings

We used two datasets, VisDrone2019 and xView [19], to evaluate our model, and we report our results using the metrics of mAP50 (the average of all ten IoU thresholds with a range of 0.5), mAP50_s ( area < 32 2 ), mAP50_m ( 32 2 area 96 2 ), and mAP50_l ( area > 96 2 ). Here, the VisDrone2019-DET dataset is the same as the VisDrone2021-DET dataset and the VisDrone2022-DET dataset.
VisDrone2019-DET [18] is an object detection dataset consisting of 8599 UAV aerial images taken at different locations and altitudes. It includes 6471 images in the training set, 1580 in the test set, and 548 in the validation set. Most of the images in this dataset have the following characteristics: (1) objects occupying small areas are densely distributed, (2) certain objects are obscured, and (3) reflections, photometric shifts, viewpoint shifts, and similar phenomena are present. The bounding boxes of more than 540,000 objects are annotated using ten predefined categories: pedestrian, person, bicycle, car, van, truck, tricycle, awning-tricycle, bus, and motor.
xView [19] is a large dataset for object detection from remote sensing images. It contains a large number of remote sensing images taken around the world, all of which have high resolutions, i.e., 3093 × 2931 . xView [19] includes more than one million object instances from 60 different categories. During the experiment, we randomly selected 90% and 10% of the images to form the training and validation sets, respectively.

4.2. Implementation Details

The training and testing of our model used a single NVIDIA RTX3080 GPU, and others were tested on a single NVIDIA RTX2080 Ti GPU. In the training phase, we used a pre-trained TPH-YOLOv5 model [10] with dataset fine-tuning for training and evaluation of the ASAHI algorithm. After using ASAF, the training set of the fine-tuning dataset contained 50,708 slicing images and 6471 original images. The sliced images were smaller than the original images. Next, we set the long side of the images to 416 pixels and the batch size to 32. We trained the model for 120 epochs, with the first two epochs used for warm-up. The Adam optimizer was used in the training process, and we used 3 × 10 3 as our initial learning rate, with the last epoch decaying to 0.12 of the initial learning rate. During inference, the overlapping was set to 0.15, the Cluster-DIoU-NMS matching threshold to 0.5, and the LS to 1818.

4.3. Ablation Studies

VisDrone2019-DET-test is a challenging dataset, containing dark scenes, multi-view shots, and varying resolutions. We illustrated the universality of our slicing method through validation on the test dataset, as detailed in Table 2. Considering the aspect ratio of the image, we tested several numbers of slices; maintaining the number of slices at 12 achieved the highest mAP50 result. However, 12 slices led to a higher computation time than a smaller number of slices. In SAHI [15], the slicing size is fixed to 512. While this results in a better performance than other sizes, a fixed slice size inevitably causes various degrees of redundant computation problems when using images with different resolutions. As shown in Table 2, we experimented with slices having a fixed size of 512 as well as fixed numbers of 4, 6, 12, and 15 slices, and found that our TPH+ASAHI (Adaptive) slicing method obtained the highest mAP50 score (45.6) and a fast inference speed (4.88), with 6 slices achieving faster computation and 12 slices showing better accuracy. In order to keep 6 and 12 patches, we set the LS (discussed in Section 3.2) to constrain the slicing number based on the image resolution. The mAP performance of ASAHI, with the specific impact of each component, is provided in Table 3. The results indicate that the proposed approach leads to improved detection performance for both small and medium objects.

4.4. Comparisons with the State of the Art

We conducted comparison experiments with the current state-of-the-art small object target detection algorithms [11,36,37,38] as well as several classical object detection algorithms [2,34]. The details are shown in Table 4. Our method reached a mAP result of 35.4 and mAP50 of 56.8. Moreover, the corresponding image processing times per second (img/s) of each components tested on VisDrone2019-DET-test are presented in Figure 8, and the discrepancy between the results and processing times of different post-processing methods, when tested on VisDrone2019-DET-val, are shown in Table 5. CDN obtained the highest mAP50 result of 48.4 and 68.6 for small and medium objects, and maintained the fastest detection speed of 5.25. The comparisons of each proposed component combined with SAHI [15] and ASAHI validated on the VisDrone2019-DET-val dataset are shown in Table 6. ASAHI performed very well on each component, especially detecting small objects, for which it obtained an mAP50 result of 56.8 and an mAP50_s result of 48.4.

5. Results and Limitations

Overall, ASAHI achieved good performance and fast inference speed on both the VisDrone2019-DET [18] and xView [19] datasets. It is worth noting that the performance on xView [19] was outstanding; this is notable because the resolution of most images in the xView [19] dataset is large, which causes significant problems for existing feature extraction algorithms. ASAHI can significantly improve this situation thanks to the proposed adaptive slicing method.
In Figure 9 and Figure 10, we exhibit a portion of the detection results of ASAHI on VisDrone and xView for small objects. As can be seen in Figure 9, ASAHI performs well in various challenging scenes, such as dark scenes (as shown in the third example), scenes with reflections and different shooting angles, and scenes with high density (as shown in the seventh example). Moreover, as shown in Figure 10, ASAHI performs equally well on remote sensing scenes, such as those with extremely large image resolutions, numerous interference factors (as shown in the first example), and multiple detection categories (as shown in the fourth example).
However, this approach has two notable shortcomings: (1) improving the detection accuracy for small objects inevitably leads to a reduction in the detection accuracy for large objects, and (2) a number of error cases remain, such as location and category confusion. The error analysis results of ASAHI on VisDrone and xView are presented on the left and right of Figure 11, respectively. Among these results, C75 (IoU threshold of 0.75), C50 (IoU threshold of 0.5), Loc (ignore location error), Sim (ignore super-category false positives), Oth (ignore category confusions), BG (ignore all false positives), and FN (ignore all false negatives) in Figure 11 clearly show that there is room to improve the error in terms of super-category false positives, category confusion, false positives, and false negatives.

6. Conclusions

In this study, we addressed the redundant computation problem of the generic SAHI slicing method [15]. We proposed a novel adaptive slicing framework, ASAHI, which slices images into a corresponding number of slices rather than using a fixed slice size. This substantially alleviates the problem of redundant computation. TPH-YOLOv5 [10] was used as the backbone to conduct relevant test experiments on the xView [19] and VisDrone2019 [18] datasets. In the post-processing stage, we replaced the traditional NMS with Cluster-NMS to improve the detection speed. Our model realizes improvements in mAP and reduces the time consumption by about 25% compared to SAHI [15] (baseline). Our method is equally applicable to other YOLO-based models. In future work, we plan to use coordinate attention [41] and CSWin Transformer [42] to improve the sensitivity of the proposed method to small objects.

Author Contributions

Conceptualization, H.Z., C.H., B.J. and B.L.; methodology, H.Z., C.H. and W.S.; software, H.Z.; validation, H.Z., C.H., W.S. and B.L.; formal analysis, H.Z.; investigation, H.Z.; resources, H.Z.; data curation, H.Z.; writing—original draft preparation, H.Z.; writing—review and editing, C.H.; visualization, H.Z.; supervision, C.H. and B.L.; project administration, C.H. and B.L.; funding acquisition, C.H. and B.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (Grant No. 61901191), National Natural Science Foundation of China (No.62177029), the Shandong Provincial Natural Science Foundation (Grant No. ZR2020LZH005), and the China Postdoctoral Science Foundation (Grant No. 2022M713668).

Data Availability Statement

The xView Dataset is available at http://xviewdataset.org/ (accessed on 22 February 2018). The VisDrone2019-DET Dataset is available at http://aiskyeye.com/download/object-detection-2/ (accessed on 3 January 2021).

Acknowledgments

The authors thank the anonymous reviewers and the editors for their insightful comments and helpful suggestions for improving our manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IoUIntersection over Union
mAPmean Average Precision
SAHISlicing-Aided Hyper Inference
NMSNon-Maximum Suppression
CDNCluster-DIoU-NMS

References

  1. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. arXiv 2017. [Google Scholar] [CrossRef]
  2. Wang, X.; Shrivastava, A.; Gupta, A.K. A-Fast-RCNN: Hard Positive Generation via Adversary for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3039–3048. [Google Scholar]
  3. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. arXiv 2015. [Google Scholar] [CrossRef] [Green Version]
  4. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. TOOD: Task-aligned One-stage Object Detection. arXiv 2021. [Google Scholar] [CrossRef]
  5. Xu, S.; Wang, X.; Lv, W.; Chang, Q.; Cui, C.; Deng, K.; Wang, G.; Dang, Q.; Wei, S.; Du, Y.; et al. PP-YOLOE: An evolved version of YOLO. arXiv 2022. [Google Scholar] [CrossRef]
  6. Qi, G.; Zhang, Y.; Wang, K.; Mazur, N.; Liu, Y.; Malaviya, D. Small Object Detection Method Based on Adaptive Spatial Parallel Convolution and Fast Multi-Scale Fusion. Remote Sens. 2022, 14, 420. [Google Scholar] [CrossRef]
  7. Pan, W.; Zhao, Z.; Huang, W.; Zhang, Z.; Fu, L.; Pan, Z.; Yu, J.; Wu, F. Video Moment Retrieval with Noisy Labels. IEEE Trans. Neural Netw. Learn. Syst. 2022. [Google Scholar] [CrossRef] [PubMed]
  8. Song, W.; Wang, X.; Liu, F. Efficient Shared Feature Learning for Cross-modality Person Re-identification. In Proceedings of the 2022 14th International Conference on Wireless Communications and Signal Processing (WCSP), Nanjing, China, 14–17 October 2022; pp. 100–105. [Google Scholar] [CrossRef]
  9. Ma, L.; Zheng, Y.; Zhang, Z.; Yao, Y.; Fan, X.; Ye, Q. Motion Stimulation for Compositional Action Recognition. IEEE Trans. Circuits Syst. Video Technol. 2022. [Google Scholar] [CrossRef]
  10. Zhu, X.; Lyu, S.; Wang, X.; Zhao, Q. TPH-YOLOv5: Improved YOLOv5 Based on Transformer Prediction Head for Object Detection on Drone-captured Scenarios. arXiv 2021. [Google Scholar] [CrossRef]
  11. Yang, C.; Huang, Z.; Wang, N. QueryDet: Cascaded Sparse Query for Accelerating High-Resolution Small Object Detection. arXiv 2021. [Google Scholar] [CrossRef]
  12. Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-NMS—Improving Object Detection with One Line of Code. arXiv 2017. [Google Scholar] [CrossRef]
  13. Solovyev, R.; Wang, W.; Gabruseva, T. Weighted boxes fusion: Ensembling boxes from different object detection models. Image Vis. Comput. 2021, 107, 104117. [Google Scholar] [CrossRef]
  14. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression. arXiv 2019. [Google Scholar] [CrossRef]
  15. Akyon, F.C.; Onur Altinuc, S.; Temizel, A. Slicing Aided Hyper Inference and Fine-Tuning for Small Object Detection. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 966–970. [Google Scholar] [CrossRef]
  16. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtually, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  17. Zheng, Z.; Wang, P.; Ren, D.; Liu, W.; Ye, R.; Hu, Q.; Zuo, W. Enhancing Geometric Factors in Model Learning and Inference for Object Detection and Instance Segmentation. arXiv 2020. [Google Scholar] [CrossRef] [PubMed]
  18. Du, D.; Zhu, P.; Wen, L.; Bian, X.; Lin, H.; Hu, Q.; Peng, T.; Zheng, J.; Wang, X.; Zhang, Y.; et al. VisDrone-DET2019: The Vision Meets Drone Object Detection in Image Challenge Results. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  19. Lam, D.; Kuzma, R.; McGee, K.; Dooley, S.; Laielli, M.; Klaric, M.; Bulatov, Y.; McCord, B. xView: Objects in Context in Overhead Imagery. arXiv 2018. [Google Scholar] [CrossRef]
  20. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018. [Google Scholar] [CrossRef]
  21. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.E.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the ECCV, Amsterdam, The Netherlands, 8–16 October 2016. [Google Scholar]
  22. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 10778–10787. [Google Scholar]
  23. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2017. [Google Scholar] [CrossRef]
  24. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. arXiv 2017. [Google Scholar] [CrossRef]
  25. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. arXiv 2013. [Google Scholar] [CrossRef]
  26. Purkait, P.; Zhao, C.; Zach, C. SPP-Net: Deep Absolute Pose Regression with Synthetic Views. arXiv 2017. [Google Scholar] [CrossRef]
  27. Qiao, S.; Chen, L.C.; Yuille, A. DetectoRS: Detecting Objects with Recursive Feature Pyramid and Switchable Atrous Convolution. arXiv 2020. [Google Scholar] [CrossRef]
  28. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. arXiv 2016. [Google Scholar] [CrossRef]
  29. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the ECCV, Munich, Germany, 8–14 September 2018. [Google Scholar]
  30. Sun, L.; Cheng, S.; Zheng, Y.; Wu, Z.; Zhang, J. SPANet: Successive Pooling Attention Network for Semantic Segmentation of Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 4045–4057. [Google Scholar] [CrossRef]
  31. Fu, L.; Zhang, D.; Ye, Q. Recurrent Thrifty Attention Network for Remote Sensing Scene Recognition. IEEE Trans. Geosci. Remote Sens. 2020, 59, 8257–8268. [Google Scholar] [CrossRef]
  32. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2020. [Google Scholar] [CrossRef]
  33. Koyun, O.C.; Keser, R.K.; Akkaya, I.B.; Töreyin, B.U. Focus-and-Detect: A Small Object Detection Framework for Aerial Images. Signal Process. Image Commun. 2022, 104, 116675. [Google Scholar] [CrossRef]
  34. Wang, J.; Xu, C.; Yang, W.; Yu, L. A Normalized Gaussian Wasserstein Distance for Tiny Object Detection. arXiv 2021. [Google Scholar] [CrossRef]
  35. Zheng, Z.; Ye, R.; Wang, P.; Ren, D.; Zuo, W.; Hou, Q.; Cheng, M.M. Localization Distillation for Dense Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 9407–9416. [Google Scholar]
  36. Deng, S.; Li, S.; Xie, K.; Song, W.; Liao, X.; Hao, A.; Qin, H. A Global-Local Self-Adaptive Network for Drone-View Object Detection. IEEE Trans. Image Process. 2021, 30, 1556–1569. [Google Scholar] [CrossRef] [PubMed]
  37. Yang, F.; Fan, H.; Chu, P.; Blasch, E.; Ling, H. Clustered Object Detection in Aerial Images. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8310–8319. [Google Scholar] [CrossRef] [Green Version]
  38. Suo, J.; Wang, T.; Zhang, X.; Chen, H.; Zhou, W.; Shi, W. HIT-UAV: A High-altitude Infrared Thermal Dataset for Unmanned Aerial Vehicles. arXiv 2022. [Google Scholar] [CrossRef]
  39. Li, C.; Yang, T.; Zhu, S.; Chen, C.; Guan, S. Density Map Guided Object Detection in Aerial Images. arXiv 2020. [Google Scholar] [CrossRef]
  40. Luo, X.; Wu, Y.; Wang, F. Target Detection Method of UAV Aerial Imagery Based on Improved YOLOv5. Remote Sens. 2022, 14, 5063. [Google Scholar] [CrossRef]
  41. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. arXiv 2021. [Google Scholar] [CrossRef]
  42. Dong, X.; Bao, J.; Chen, D.; Zhang, W.; Yu, N.; Yuan, L.; Chen, D.; Guo, B. CSWin Transformer: A General Vision Transformer Backbone with Cross-Shaped Windows. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 12124–12134. [Google Scholar]
Figure 1. Inference results for (a) YOLOV5, (b) TPH-YOLOV5, (c) TPH-YOLOV5+SAHI, and (d) TPH-YOLOV5+ASAHI (our algorithm). As shown in the magenta boxes, in (c,d) more bounding boxes are detected compared with (a,b). In addition, while the slicing method from [15] performs well when detecting small objects, ASAHI is superior when detecting even smaller objects. This is evident in the green closeups, where ASAHI is able to detect more tiny objects.
Figure 1. Inference results for (a) YOLOV5, (b) TPH-YOLOV5, (c) TPH-YOLOV5+SAHI, and (d) TPH-YOLOV5+ASAHI (our algorithm). As shown in the magenta boxes, in (c,d) more bounding boxes are detected compared with (a,b). In addition, while the slicing method from [15] performs well when detecting small objects, ASAHI is superior when detecting even smaller objects. This is evident in the green closeups, where ASAHI is able to detect more tiny objects.
Remotesensing 15 01249 g001
Figure 2. The overall flow chart of ASAHI. (1) Based on LS (the threshold image resolution for different slicing methods), images with different resolutions are sliced into 6 and 12 patches, as shown in (A). (2) During the inference stage, the original image and the sliced image are sent to TPH-YOLOV5 [10] to extract the image features, as shown in (B), and to infer the bounding boxes. (3) The original image and sliced image are joined together by TPH-YOLOV5 [10] to infer the bounding boxes. Finally, all bounding boxes are aggregated by Cluster-DIoU-NMS to suppress redundant bounding boxes, as shown in (C).
Figure 2. The overall flow chart of ASAHI. (1) Based on LS (the threshold image resolution for different slicing methods), images with different resolutions are sliced into 6 and 12 patches, as shown in (A). (2) During the inference stage, the original image and the sliced image are sent to TPH-YOLOV5 [10] to extract the image features, as shown in (B), and to infer the bounding boxes. (3) The original image and sliced image are joined together by TPH-YOLOV5 [10] to infer the bounding boxes. Finally, all bounding boxes are aggregated by Cluster-DIoU-NMS to suppress redundant bounding boxes, as shown in (C).
Remotesensing 15 01249 g002
Figure 3. The architecture of TPH-YOLOV5 [10] with slight changes made for this study. (1) The backbone is based on CSPDarknet, with three added transformer encoder blocks, each of which contains two main blocks, a multi-head attention block, and a feed forward neural network at the end. (2) The neck uses the structure of the PANet and is optimized by using CBAM [29]. The entire model diagram is based on TPH-YOLOV5 [10].
Figure 3. The architecture of TPH-YOLOV5 [10] with slight changes made for this study. (1) The backbone is based on CSPDarknet, with three added transformer encoder blocks, each of which contains two main blocks, a multi-head attention block, and a feed forward neural network at the end. (2) The neck uses the structure of the PANet and is optimized by using CBAM [29]. The entire model diagram is based on TPH-YOLOV5 [10].
Remotesensing 15 01249 g003
Figure 4. In the inference stage, the image is divided into two parts, namely, FI and ASAHI. These processes can be carried out simultaneously or sequentially. In ASAHI, the image is divided into several small patches according to the image resolution, and predictions are generated from these resized patches. Next, the results from ASAHI and FI are merged by Cluster-DIoU-NMS and converted back into the original image.
Figure 4. In the inference stage, the image is divided into two parts, namely, FI and ASAHI. These processes can be carried out simultaneously or sequentially. In ASAHI, the image is divided into several small patches according to the image resolution, and predictions are generated from these resized patches. Next, the results from ASAHI and FI are merged by Cluster-DIoU-NMS and converted back into the original image.
Remotesensing 15 01249 g004
Figure 5. The redundant computation problem addressed by SAHI [15].
Figure 5. The redundant computation problem addressed by SAHI [15].
Remotesensing 15 01249 g005
Figure 6. Schematic diagram of traditional NMS and Cluster-NMS. In traditional NMS, b is obtained from X that has been processed after the maximization, then the binarization operation is conducted according to NMS threshold, as shown on the left. Based on the Traditional NMS, Cluster-NMS tackles the oversuppression problem by left-multiplying b, as shown on the right.
Figure 6. Schematic diagram of traditional NMS and Cluster-NMS. In traditional NMS, b is obtained from X that has been processed after the maximization, then the binarization operation is conducted according to NMS threshold, as shown on the left. Based on the Traditional NMS, Cluster-NMS tackles the oversuppression problem by left-multiplying b, as shown on the right.
Remotesensing 15 01249 g006
Figure 7. Flowchart of ASAF. The pre-training dataset is sliced to form the slicing dataset using ASAHI, then the fine-tuned dataset is set up together with the pre-training dataset.
Figure 7. Flowchart of ASAF. The pre-training dataset is sliced to form the slicing dataset using ASAHI, then the fine-tuned dataset is set up together with the pre-training dataset.
Remotesensing 15 01249 g007
Figure 8. Comparison of image processing speeds between SAHI [15] and ASAHI under different component conditions. The corresponding colors are used to represent comparison of the same component. After calculating the average image processing speed over five experiments on VisDrone2019-DET-test, we found that it increased by 0.48 img/s with ASAHI, 0.97 img/s with ASAHI and PO, and 1.19 img/s with ASAHI, PO, and CDN. Although FI and ASAF cannot significantly improve detection speed, they can enhance detection performance.
Figure 8. Comparison of image processing speeds between SAHI [15] and ASAHI under different component conditions. The corresponding colors are used to represent comparison of the same component. After calculating the average image processing speed over five experiments on VisDrone2019-DET-test, we found that it increased by 0.48 img/s with ASAHI, 0.97 img/s with ASAHI and PO, and 1.19 img/s with ASAHI, PO, and CDN. Although FI and ASAF cannot significantly improve detection speed, they can enhance detection performance.
Remotesensing 15 01249 g008
Figure 9. Example detection results of ASAHI on the VisDrone2019-DET-val dataset.
Figure 9. Example detection results of ASAHI on the VisDrone2019-DET-val dataset.
Remotesensing 15 01249 g009
Figure 10. Example detection results of ASAHI on the xView-val dataset.
Figure 10. Example detection results of ASAHI on the xView-val dataset.
Remotesensing 15 01249 g010
Figure 11. Error analysis curve for ASAHI testing on all classes in the VisDrone2019-DET-val set (left) and error analysis bar plot for ASAHI testing on the vehicles class in the xView-val set (right). In the error analysis curve (left), category confusion is the main cause of errors, at 11.4 (Oth minus C50). In contrast, in the error analysis bar plot (right) location errors are the main source of error, occurring in 14% of cases.
Figure 11. Error analysis curve for ASAHI testing on all classes in the VisDrone2019-DET-val set (left) and error analysis bar plot for ASAHI testing on the vehicles class in the xView-val set (right). In the error analysis curve (left), category confusion is the main cause of errors, at 11.4 (Oth minus C50). In contrast, in the error analysis bar plot (right) location errors are the main source of error, occurring in 14% of cases.
Remotesensing 15 01249 g011
Table 1. Reduction in the redundancy of the computed area with ASAHI as compared to SAHI [15] at different resolutions. All image resolutions are from the VisDrone and xView [19] datasets.
Table 1. Reduction in the redundancy of the computed area with ASAHI as compared to SAHI [15] at different resolutions. All image resolutions are from the VisDrone and xView [19] datasets.
Image ResolutionRedundant Calculation Reduction (%)
960 × 540 38.72
1360 × 765 2.56
1400 × 1050 25.61
1920 × 1080 25.92
2000 × 1500 24.13
2913 × 2428 6.99
Table 2. Results of testing the proposed ASAHI method with different slice sizes and the baseline SAHI method with a fixed slice size of 512 on the VisDrone2019-DET-test dataset. The best results are indicated in bold.
Table 2. Results of testing the proposed ASAHI method with different slice sizes and the baseline SAHI method with a fixed slice size of 512 on the VisDrone2019-DET-test dataset. The best results are indicated in bold.
Set Upimg/smAP50mAP50_smAP50_mmAP50_l
TPH + ASAHI (4 slice)5.1938.724.956.568.8
TPH + ASAHI (6 slice)4.9841.628.858.565.3
TPH + ASAHI (12 slice)2.9841.930.556.464.3
TPH + ASAHI (15 slice)2.3940.929.955.162.9
TPH + SAHI (512) (baseline)3.6942.229.858.565.6
TPH + ASAHI (Adaptive)4.88 (↑1.19)45.6 (↑3.4)33.8 (↑4.0)59.7 (↑1.2)60.7
Table 3. The importance of each proposed component validated on VisDrone2019-DET-val. Here, PO and CDN stand for overlapping patches and Cluster-DIoU-NMS, respectively. The best results are indicated in bold.
Table 3. The importance of each proposed component validated on VisDrone2019-DET-val. Here, PO and CDN stand for overlapping patches and Cluster-DIoU-NMS, respectively. The best results are indicated in bold.
ModelmAP50mAP50_smAP50_mmAP50_l
TPH + FI34.221.453.374.4
TPH + ASAHI54.446.666.158.9
TPH + ASAHI + FI55.546.768.176.5
TPH + ASAHI + FI + PO55.546.868.677.6
TPH + ASAHI + FI + PO + ASAF + CDN56.848.468.672.9
Table 4. Comparison of our method with other state-of-the-art object detection methods on VisDrone2019-DET-val. Certain results are taken from other studies, as indicated. The best results are indicated in bold.
Table 4. Comparison of our method with other state-of-the-art object detection methods on VisDrone2019-DET-val. Certain results are taken from other studies, as indicated. The best results are indicated in bold.
ModelmAPmAP50img/s
Fast-RCNN [2]-38.5-
Cascade R-CNN + NWD [34]-40.3-
DMNet [39]28.249.3-
YOLO-UAV [40]30.5--
HIT-UAV [38]-55.8-
CRENet33.754.31.10
GLSAN [36]30.755.4-
QueryDet [11]33.956.12.75
ClusDet [37]32.456.21.29
Focus-And-Detect [33]42.066.10.73
TPH + SAHI34.955.15.01
TPH + ASAHI(ours)35.456.84.88
Table 5. The mAP and time consumption results obtained with different post-processing methods under ASAHI on VisDrone2019-DET-val. The best results are indicated in bold.
Table 5. The mAP and time consumption results obtained with different post-processing methods under ASAHI on VisDrone2019-DET-val. The best results are indicated in bold.
ModelmAP50mAP50_smAP50_mmAP50_limg/s
NMS55.847.468.372.93.06
SOFT-NMS [12]52.644.365.870.02.72
WBF [13]55.146.767.573.32.88
Cluster-GIoU-NMS56.348.068.572.95.17
Cluster-CIoU-NMS56.848.468.672.94.87
Cluster-DIoU-NMS [35]56.848.468.672.95.26
Table 6. The importance of each component as validated on VisDrone2019-DET-val. The best results are indicated in bold.
Table 6. The importance of each component as validated on VisDrone2019-DET-val. The best results are indicated in bold.
ModelmAP50mAP50_smAP50_mmAP50_l
TPH + FI34.221.453.374.4
TPH + SAHI54.245.766.869.0
TPH + SAHI + PO54.846.267.867.2
TPH + SAHI + FI54.645.768.375.2
TPH + SAHI + FI + PO55.146.168.874.6
TPH + SAHI + FI + PO + ASAF + CDN56.548.169.072.7
TPH + ASAHI54.446.666.158.9
TPH + ASAHI + FI55.546.768.176.5
TPH + ASAHI + FI + PO55.546.868.677.6
TPH + ASAHI + FI + PO + ASAF + CDN56.848.468.672.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Hao, C.; Song, W.; Jiang, B.; Li, B. Adaptive Slicing-Aided Hyper Inference for Small Object Detection in High-Resolution Remote Sensing Images. Remote Sens. 2023, 15, 1249. https://doi.org/10.3390/rs15051249

AMA Style

Zhang H, Hao C, Song W, Jiang B, Li B. Adaptive Slicing-Aided Hyper Inference for Small Object Detection in High-Resolution Remote Sensing Images. Remote Sensing. 2023; 15(5):1249. https://doi.org/10.3390/rs15051249

Chicago/Turabian Style

Zhang, Hao, Chuanyan Hao, Wanru Song, Bo Jiang, and Baozhu Li. 2023. "Adaptive Slicing-Aided Hyper Inference for Small Object Detection in High-Resolution Remote Sensing Images" Remote Sensing 15, no. 5: 1249. https://doi.org/10.3390/rs15051249

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop