Next Article in Journal
Low-Rank Constrained Attention-Enhanced Multiple Spatial–Spectral Feature Fusion for Small Sample Hyperspectral Image Classification
Previous Article in Journal
Sudden Track Turning of Typhoon Prapiroon (2012) Enhanced the Upper Ocean Response
Previous Article in Special Issue
Deep Learning Approach for Object Classification on Raw and Reconstructed GBSAR Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Editorial

Synthetic Aperture Radar (SAR) Meets Deep Learning

1
School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2
Electrical and Electronic Engineering, University of Hong Kong, Hong Kong 999077, China
3
School of Aeronautics and Astronautics, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(2), 303; https://doi.org/10.3390/rs15020303
Submission received: 6 December 2022 / Accepted: 3 January 2023 / Published: 4 January 2023
(This article belongs to the Special Issue Synthetic Aperture Radar (SAR) Meets Deep Learning)

1. Introduction

Synthetic aperture radar (SAR) is an important active microwave imaging sensor. Its all-day and all-weather working capacity makes it play an important role in the remote sensing community. Since the launch of the first SAR satellite by the United States [1], SAR has received extensive attention in the remote sensing community [2], e.g., geological exploration [3], topographic mapping [4], disaster forecast [5,6], and marine traffic management [7,8,9,10]. Therefore, it is valuable and meaningful to study SAR-based remote sensing applications [11].
In recent years, with the rapid development of artificial intelligence, deep learning (DL) [12] has been applied to all walks of life, such as face recognition, automatic driving, search recommendation, internet of things, and so on. The DL represented by convolutional neural network (CNN) is promoting the evolution of many algorithms and the innovation of advanced technologies. At present, scholars are exploring the application value of DL in SAR remote sensing field. Many SAR remote sensing application technologies based on DL have emerged, such as land surface change detection, ocean remote sensing, sea-land segmentation, traffic surveillance and topographic mapping.
Aiming to promote the application of DL in SAR, we initiated this Special Issue and collected a total of 14 papers (including 12 articles, 1 review and 1 technical note) covering various topics, e.g., object detection, classification and tracking, SAR image intelligent processing, data analytics in the SAR remote sensing community and interferometric SAR technology. The overview of contribution is in the following section.

2. Overview of Contribution

On the topic of object detection, classification and tracking, Li et al. [13] summarized the dataset, algorithm, performance, DL framework, country and timeline of DL-based ship detection methods. They analyzed the 177 published papers about DL-based SAR ship detection and attempted to stimulate more research in this field. Xia [14] proposed a visual transformer framework based on contextual joint-representation learning referred to as CRTransSar. CRTransSar combined the global contextual information perception of transformers and the local feature representation capabilities of convolutional neural networks (CNNs). It was found to produce more accurate ship detection results than other most advanced methods. Note that the authors also released a larger-scale SAR multiclass target detection dataset called SMCDD. Feng et al. [15] established a lightweight position-enhanced anchor-free SAR ship detection algorithm called LPEDet. They designed a lightweight multiscale backbone and a position-enhanced attention strategy for balancing detection speed and accuracy. The results showed that their method achieved a higher detection accuracy and a faster detection speed than other state-of-the-art (SOTA) detection methods. Xu et al. [16] presented a unified framework combining triangle distance IoU loss (TDIoU loss), an attention-weighted feature pyramid network (AW-FPN), and a Rotated-SARShip dataset (RSSD) for arbitrary-oriented SAR ship detection. Their method showed superior performance on both SAR and optical image datasets, significantly outperforming the SOTA methods. Xiao et al. [17] proposed a simple, yet effective, self-supervised representation learning (Lite-SRL) algorithm for the scene classification task. Note that they successfully evaluate the on-board operational capability of Lite-SRL by transplanting Lite-SRL to the low-power computing platform NVIDIA Jetson TX2. Kačan et al. [18] explored object classification on a raw and a reconstructed Ground-based SAR (GBSAR) data. They revealed how processing raw data provides overall better classification accuracy than processing reconstructed data, and revealed the value of this method in industrial GBSAR applications where processing speed is critical. Bao et al. [19] proposed a guided anchor Siamese network (GASN) for arbitrary targets of interest (TOI) tracking in Video-SAR. GASN used a matching function for returning the most similar area, followed by a guided anchor subnetwork to suppress false alarms. GASN realized the TOI tracking with high diversity and arbitrariness, outperforming SOTA methods.
On the topic of SAR image intelligent processing, Tan et al. [20] proposed a feature-preserving heterogeneous remote sensing image transformation model. Through decoupling network design, the method enabled enhancing the detailed information of the generated optical images and reducing its spectral distortion. The results in SEN-2 satellite images revealed that the proposed model has obvious advantages in feature reconstruction and the economical volume of the parameters. Zhang et al. [21] proposed a self-supervised despeckling algorithm with an enhanced U-Net called SSEUNet. Unlike previous self-supervised despeckling works, the noisy-noisy image pairs in SSEUNet were generated from real-world SAR images through a novel generation training pairs module, making it possible to train deep convolutional neural networks using real-world SAR images. Finally, experiments on simulated and real-world SAR images show that SSEUNet notably exceeds SOTA despeckling methods. Habibollahi et al. [22] proposed a DL-based change detection algorithm for bi-temporal polarimetric SAR (PolSAR) imagery called TCD-Net. In particular, this method applied three steps as follows: (1) pre-processing, (2) parallel pseudo-label training sample generation based on a pre-trained model and the fuzzy C-means (FCM) clustering algorithm, and (3) classification. TCD-Net could learn more strong and abstract representations for the spatial information of a certain pixel, and was superior to other well-known methods. Fan et al. [23] proposed a high-precision, rapid, large-size SAR image dense-matching method. The method mainly included four steps: down-sampling image pre-registration, sub-image acquisition, dense matching, and the transformation solution. The experimental results demonstrated that the proposed method is efficient and accurate, which provides a new idea for SAR image registration. Zhang et al. [24] proposed A low-grade road extraction network Based on the fusion of optical and SAR images at the decision level called SDG-DenseNet. Furthermore, they verified that the decision-level fusion of road binary maps from SAR and optical images can significantly improve the accuracy of low-grade road extraction from remote sensing images.
On the topic of data analytics in the SAR remote sensing community, Wangiyana et al. [25] explored the impact of several data augmentation (DA) methods on the performance of building detection on a limited dataset of SAR images. Their results showed that geometric transformations are more effective than pixel transformations and DA methods should be used in moderation to prevent unwanted transformations outside the possible object variations. The study could provide potential guidelines for future research in selecting DA methods for segmentation tasks in radar imagery.
On the topic of interferometric SAR technology, Pu et al. [26] proposed a robust least squares phase unwrapping method called PGENet that works via a phase gradient estimation network based on the encoder–decoder architecture for InSAR. Experiments on simulated and real InSAR data demonstrated that PGENet outperformed the other five well-established phase unwrapping methods and was robust to noise.

3. Conclusions

Recently, as many SAR systems have been put into use, massive SAR data are available, providing important support for exploring how to apply DL to SAR fields. A large number of SAR data coupled with the DL methodology jointly promote the development of SAR fields. The Special Issue shows innovative applications in object detection, classification and tracking, SAR image intelligent processing, data analytics in the SAR remote sensing community and interferometric SAR technology. There is no doubt that applying DL to more SAR fields (such as terrain classification, SAR agriculture monitoring, SAR imaging algorithm updating, SAR forest applications, marine pollution, etc.) is of great significance for earth remote sensing. In addition, we welcome scholars who are interested in applying DL to SAR to contribute to the scientific literature on this subject.

Author Contributions

Writing—original draft preparation, T.Z. (Tianwen Zhang); writing—review and editing, T.Z. (Tiaojiao Zeng) and X.Z. All authors have read and agreed to the published version of the manuscript.

Data Availability Statement

Not applicable.

Acknowledgments

We thank all authors, reviewers and editors for their contributions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Born, G.H.; Dunne, J.A.; Lame, D.B. Seasat mission overview. Science 1979, 204, 1405–1406. [Google Scholar] [CrossRef]
  2. Moreira, A.; Prats-Iraola, P.; Younis, M.; Krieger, G.; Hajnsek, I.; Papathanassiou, K.P. A tutorial on synthetic aperture radar. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–43. [Google Scholar] [CrossRef] [Green Version]
  3. De Novellis, V.; Castaldo, R.; Lollino, P.; Manunta, M.; Tizzani, P. Advanced Three-Dimensional Finite Element Modeling of a Slow Landslide through the Exploitation of DInSAR Measurements and in Situ Surveys. Remote Sens. 2016, 8, 670. [Google Scholar] [CrossRef] [Green Version]
  4. Da Silva, A.D.Q.; Paradella, W.R.; Freitas, C.C.; Oliveira, C.G. Evaluation of Digital Classification of Polarimetric SAR Data for Iron-Mineralized Laterites Mapping in the Amazon Region. Remote Sens. 2013, 5, 3101–3122. [Google Scholar] [CrossRef] [Green Version]
  5. Khan, S.I.; Hong, Y.; Gourley, J.J.; Khattak, M.U.; De Groeve, T. Multi-Sensor Imaging and Space-Ground Cross-Validation for 2010 Flood along Indus River, Pakistan. Remote Sens. 2014, 6, 2393–2407. [Google Scholar] [CrossRef] [Green Version]
  6. Martinis, S.; Twele, A.; Strobl, C.; Kersten, J.; Stein, E. A Multi-Scale Flood Monitoring System Based on Fully Automatic MODIS and TerraSAR-X Processing Chains. Remote Sens. 2013, 5, 5598–5619. [Google Scholar] [CrossRef] [Green Version]
  7. Xu, X.; Zhang, X.; Zhang, T. Lite-YOLOv5: A Lightweight Deep Learning Detector for On-Board Ship Detection in Large-Scene Sentinel-1 SAR Images. Remote Sens. 2022, 14, 1018. [Google Scholar] [CrossRef]
  8. Zhang, T.; Zhang, X. A Mask Attention Interaction and Scale Enhancement Network for SAR Ship Instance Segmentation. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4511005. [Google Scholar] [CrossRef]
  9. Xu, X.; Zhang, X.; Shao, Z.; Shi, J.; Wei, S.; Zhang, T.; Zeng, T. A Group-Wise Feature Enhancement-and-Fusion Network with Dual-Polarization Feature Enrichment for SAR Ship Detection. Remote Sens. 2022, 14, 5276. [Google Scholar] [CrossRef]
  10. Zhang, T.; Zhang, X. HTC+ for SAR Ship Instance Segmentation. Remote Sens. 2022, 14, 2395. [Google Scholar] [CrossRef]
  11. Zhang, L.; Zhang, L. Artificial Intelligence for Remote Sensing Data Analysis: A review of challenges and opportunities. IEEE Geosci. Remote Sens. Mag. 2022, 10, 270–294. [Google Scholar] [CrossRef]
  12. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  13. Li, J.; Xu, C.; Su, H.; Gao, L.; Wang, T. Deep Learning for SAR Ship Detection: Past, Present and Future. Remote Sens. 2022, 14, 2712. [Google Scholar] [CrossRef]
  14. Xia, R.; Chen, J.; Huang, Z.; Wan, H.; Wu, B.; Sun, L.; Yao, B.; Xiang, H.; Xing, M. CRTransSar: A Visual Transformer Based on Contextual Joint Representation Learning for SAR Ship Detection. Remote Sens. 2022, 14, 1488. [Google Scholar] [CrossRef]
  15. Feng, Y.; Chen, J.; Huang, Z.; Wan, H.; Xia, R.; Wu, B.; Sun, L.; Xing, M. A Lightweight Position-Enhanced Anchor-Free Algorithm for SAR Ship Detection. Remote Sens. 2022, 14, 1908. [Google Scholar] [CrossRef]
  16. Xu, Z.; Gao, R.; Huang, K.; Xu, Q. Triangle Distance IoU Loss, Attention-Weighted Feature Pyramid Network, and Rotated-SARShip Dataset for Arbitrary-Oriented SAR Ship Detection. Remote Sens. 2022, 14, 4676. [Google Scholar] [CrossRef]
  17. Xiao, X.; Li, C.; Lei, Y. A Lightweight Self-Supervised Representation Learning Algorithm for Scene Classification in Spaceborne SAR and Optical Images. Remote Sens. 2022, 14, 2956. [Google Scholar] [CrossRef]
  18. Kačan, M.; Turčinović, F.; Bojanjac, D.; Bosiljevac, M. Deep Learning Approach for Object Classification on Raw and Reconstructed GBSAR Data. Remote Sens. 2022, 14, 5673. [Google Scholar] [CrossRef]
  19. Bao, J.; Zhang, X.; Zhang, T.; Shi, J.; Wei, S. A Novel Guided Anchor Siamese Network for Arbitrary Target-of-Interest Tracking in Video-SAR. Remote Sens. 2021, 13, 4504. [Google Scholar] [CrossRef]
  20. Tan, D.; Liu, Y.; Li, G.; Yao, L.; Sun, S.; He, Y. Serial GANs: A Feature-Preserving Heterogeneous Remote Sensing Image Transformation Model. Remote Sens. 2021, 13, 3968. [Google Scholar] [CrossRef]
  21. Zhang, G.; Li, Z.; Li, X.; Liu, S. Self-Supervised Despeckling Algorithm with an Enhanced U-Net for Synthetic Aperture Radar Images. Remote Sens. 2021, 13, 4383. [Google Scholar] [CrossRef]
  22. Habibollahi, R.; Seydi, S.T.; Hasanlou, M.; Mahdianpari, M. TCD-Net: A Novel Deep Learning Framework for Fully Polarimetric Change Detection Using Transfer Learning. Remote Sens. 2022, 14, 438. [Google Scholar] [CrossRef]
  23. Fan, Y.; Wang, F.; Wang, H. A Transformer-Based Coarse-to-Fine Wide-Swath SAR Image Registration Method under Weak Texture Conditions. Remote Sens. 2022, 14, 1175. [Google Scholar] [CrossRef]
  24. Zhang, J.; Li, Y.; Si, Y.; Peng, B.; Xiao, F.; Luo, S.; He, L. A Low-Grade Road Extraction Method Using SDG-DenseNet Based on the Fusion of Optical and SAR Images at Decision Level. Remote Sens. 2022, 14, 2870. [Google Scholar] [CrossRef]
  25. Wangiyana, S.; Samczyński, P.; Gromek, A. Data Augmentation for Building Footprint Segmentation in SAR Images: An Empirical Study. Remote Sens. 2022, 14, 2012. [Google Scholar] [CrossRef]
  26. Pu, L.; Zhang, X.; Zhou, Z.; Li, L.; Zhou, L.; Shi, J.; Wei, S. A Robust InSAR Phase Unwrapping Method via Phase Gradient Estimation Network. Remote Sens. 2021, 13, 4564. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, T.; Zeng, T.; Zhang, X. Synthetic Aperture Radar (SAR) Meets Deep Learning. Remote Sens. 2023, 15, 303. https://doi.org/10.3390/rs15020303

AMA Style

Zhang T, Zeng T, Zhang X. Synthetic Aperture Radar (SAR) Meets Deep Learning. Remote Sensing. 2023; 15(2):303. https://doi.org/10.3390/rs15020303

Chicago/Turabian Style

Zhang, Tianwen, Tianjiao Zeng, and Xiaoling Zhang. 2023. "Synthetic Aperture Radar (SAR) Meets Deep Learning" Remote Sensing 15, no. 2: 303. https://doi.org/10.3390/rs15020303

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop