Next Article in Journal
Testing Thermostatic Bath End-Scale Stability for Calibration Performance with a Multiple-Sensor Ensemble Using ARIMA, Temporal Stochastics and a Quantum Walker Algorithm
Next Article in Special Issue
Reparameterizable Multibranch Bottleneck Network for Lightweight Image Super-Resolution
Previous Article in Journal
Effects of Charge Traps on Hysteresis in Organic Field-Effect Transistors and Their Charge Trap Cause Analysis through Causal Inference Techniques
Previous Article in Special Issue
Blind Watermarking for Hiding Color Images in Color Images with Super-Resolution Enhancement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SAR Image Ship Target Detection Adversarial Attack and Defence Generalization Research

1
Department of Information and Communication Engineering, School of Electronic Information Engineering, East Campus of Changchun University of Science and Technology, 7089 Weixing Road, Changchun 130022, China
2
Department of Robotics, School of Electronic Information Engineering, East Campus of Changchun University, 6543 Weixing Road, Changchun 130022, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 2266; https://doi.org/10.3390/s23042266
Submission received: 20 December 2022 / Revised: 13 February 2023 / Accepted: 15 February 2023 / Published: 17 February 2023
(This article belongs to the Special Issue Image Denoising and Image Super-resolution for Sensing Application)

Abstract

:
The synthetic aperture radar (SAR) image ship detection system needs to adapt to an increasingly complicated actual environment, and the requirements for the stability of the detection system continue to increase. Adversarial attacks deliberately add subtle interference to input samples and cause models to have high confidence in output errors. There are potential risks in a system, and input data that contain confrontation samples can be easily used by malicious people to attack the system. For a safe and stable model, attack algorithms need to be studied. The goal of traditional attack algorithms is to destroy models. When defending against attack samples, a system does not consider the generalization ability of the model. Therefore, this paper introduces an attack algorithm which can improve the generalization of models by based on the attributes of Gaussian noise, which is widespread in actual SAR systems. The attack data generated by this method have a strong effect on SAR ship detection models and can greatly reduce the accuracy of ship recognition models. While defending against attacks, filtering attack data can effectively improve the model defence capabilities. Defence training greatly improves the anti-attack capacity, and the generalization capacity of the model is improved accordingly.

1. Introduction

Synthetic aperture radar (SAR) [1,2] is a high-resolution imaging radar developed with digital processing technology. It has good all-weather capabilities and perspective and has been widely used in military reconnaissance surveying and mapping. Many researchers had studied the application of remote sensing in the ocean [3,4,5] for sensing the marine environment, marine and coastal environment detection. SAR was widely used in marine monitoring: oil spill monitoring [6] and marine biodiversity observation [7]. Ships are necessary equipment for marine development, energy transportation, national defence construction and other activities. The most important marine activity component is ship monitoring. SAR can penetrate clouds and fog with unique advantages and is not limited by meteorological conditions. It is very suitable for data sources for ship detection.
A ground SAR detection system operates in space tracks. Researchers have developed high-performance deep learning algorithms and frameworks. They are widely used in SAR ship detection systems to solve practical problems. Although satellite communication has a high technical level, it is still vulnerable to security threats, which may seriously affect the judgement of the signal receiver. While the detection accuracy is important, we are more concerned with the relevant security risks of models. Neural networks have reached human performance on independent identically distributed testing machines [8,9,10]. However, it is difficult to distinguish an image that contains weak naked noise that the eyes cannot detect. This noisy image data, which have serious impacts on automatic detection systems, cause a detected object to deviate from the real object or a significant offset in the detection of anchor frames. Confrontation data [11] are slightly modified images. The purpose of the modification is to interfere with the result of a machine learning analyser. The model parameters can be modified by modifying the attack data without modifying the program. This attack will have a serious impact on a model, and it is very easy to cause misdetections, missed reports or false reports. SAR image ship-detection systems need to adapt to an increasingly complicated actual environment, and the requirements for the stability of detection systems continue to increase. The focus of this research is to ensure the stability of a system, prevent the system from being disturbed by external noise, and prevent a decrease in accuracy of a detection system when there is minor interference.
This article answers the following questions:
  • What is the impact of protecting SAR ship models from attacks?
  • How can stronger confrontation data be generated?
  • Which attack form has the greatest impact on a SAR ship detection model?
  • How does these model samples help the model improve the accuracy in the defence?
The structure of the paper is as follows: The second part introduces related work, and the third part studies the expansive adaptive gradient as well as its derivative formula. The fourth part evaluates the gradient expansion attack model. The fifth part is the conclusion.

2. Related Work

2.1. Adversarial Attack

The use of machine learning in smart networks brings potential security threats. The purpose of maliciously injected fake training data is to destroy the learning model. Due to the development of neural network adversarial attacks, researchers have conducted much research on offensive attacks. According to the degree of model consistency, they can be divided into black box attacks and white box attacks. Black box attacks [12,13] can only launch confrontation attacks by querying the output classification results of an input sample. White box attacks [14,15] use information, such as model structures, parameters and other information, to conduct confrontation attacks. Common attack methods include gradient information algorithms and interpreted algorithms based on neural networks. The target data set of a SAR image ship model is also easily affected by sample attacks. Wang et al. [16] applied the Momentum Iterative Fast Gradient Sign Method (MI-FGSM) and ADVGAN algorithms to SAR data sets to generate confrontation samples and conduct SAR image classification attacks. The experimental results show that confrontation samples are destructive for SAR image models.

2.2. Gradient-Based Attack Methods

In machine learning algorithms, when minimizing the loss function, a minimum loss function and the corresponding parameter values can be found through gradient decline. Conversely, to maximize the loss function, it can be found through gradient expansion. By disturbing a vector, it can be superimposed on a sample to form an attack sample. Through an attack, the output results are as large as possible with the deviation in the input.
In this section, we briefly introduce several gradient-based attack methods. A white-box attack is based on gradient-based optimization and restrained optimization. Among them, several types of gradient optimization have a greater impact.
The Fast Gradient Sign Method (FGSM) [13] is a basic method for white box attacks. This method quickly guides the model to find the direction in which the loss function increases:
x = x + · s i g n ( x J ( f ( x ) , y ) )
The Basic Iterative Method (BIM) [17] is one of many extensions of FGSM. A confrontational example generated by BIM is defined as follows:
x i + 1 = x i + α · s i g n ( x J ( f ( x i ) , y ) )
where x 0 = x , α = ϵ / T , and T denotes the number of iterations.
TIM [18] uses Gaussian nuclear volume stairs and can be combined with MIM:
g i + 1 = μ · g i + W x J ( f ( x i ) , y ) W x J ( f ( x i ) , y ) 1
x i + 1 C l i p ϵ { x i + α · s i g n ( g i + 1 ) }
where W is the Gaussian kernel and * indicates the convolutional operator.
Dong et al. [19] proposed Momentum Iterative Fast Gradient Sign Method (MI-FGSM) and Nesterov Iterative Fast Gradient Sign Method (NI-FGSM). MI-FGSM integrates momentum into iterative attacks, thereby providing a higher transitability for confrontation examples. NI-FGSM improves the transplantability of confrontation examples. A Nesterov accelerated gradient can be integrated into an iterative gradient [20] basic attack to obtain a robust attack model.
This simple method is used to illustrate the success rate of using iteration methods in subsequent work. To generate malicious data that may be classified by a model, more iterative updates are needed. The calculation time and the number of iterations are linear, so more time is needed to create stronger attacks.
Later, the predictable perturbation attack (PPA) algorithm was used to add restricted indicators as the gradient rises.

2.3. Adversarial Defence

Studies have shown that deep learning networks have obvious weaknesses when processing data that contain noise. The disturbances contained in these pictures are intentional, very small, and cannot be perceived by human beings. This noise is difficult to detect by human eyes, but it has a serious impact on the test results. To solve this problem, researchers have designed many defence methods that focus on using good models trained on large data sets to correct adversarial examples. These methods include the adversarial training method [21,22,23], gradient regularization method [24] and method based on input transformations [25,26,27].
For example, in [28], the DeepFool algorithm was proposed to effectively calculate the disturbance cause by a deceptive deep network, thereby reliably quantifying the robustness of classifiers. A large number of experiments show that this method is better than existing methods in calculating improving a classifier’s robustness.

2.4. SAR Image Noise

SAR image noise is mainly of two types: additive noise and multiplication noise. This article focuses on the impact of additive noise on a SAR ship detection system. Additive noise usually uses zero average white noise as a model [29]. In the subsequent sections, we use the noise model that is the closest to the actual noise distribution, that is, normal distribution noise with a mean value of 0, to protect against attack noise.

3. Materials and Methods

3.1. Defending SAR Ship Data (NAA)

After a ship’s information is utilized, if the system is attacked, the hidden hazard information will bypass the defence model to attack the system, which will cause the ship’s detection model to fail and affect the test results.
There are three general approaches to defend against attacks during the model training phase: adjusting the training data, adjusting the labels, and adjusting the input features. It is more difficult for an attacker to attack model labels and input features. Adjusting the training data is the most obvious and most effective attack-response method. Hence, this article proposes a ship detection data attack model. Through these methods, the original distribution of the training data is attacked by injecting confrontation samples and adjusting, modifying or deleting training data. The adjusted data will cause the model to misjudge the detection results to achieve an attack effect.
The framework of this article consists of two parts: sensitive directional estimation and disturbance selection.

3.2. Sensitive Directional Estimation

Sensitive directional estimation refers to the opponent’s sensitivity to changes in the altered features of each input feature through the data stream around sample X.

3.3. Disturbance Signal Selection

Interference selection refers to the use of the characteristics of the information of a ship to select “interference” σ signals to obtain the most effective confrontation results. At the beginning of each new iteration, the modified sample σ + X replaces the original pure input sample X until the disturbance sample meets the opposition target conditions. Therefore, the goal is to determine a suitable disturbance. The total interference added to the original sample is as small as possible to achieve the goal of attacking the model while not being detected by the human eye.
The specific design ideas are as follows:
X is the input sample, and f is the model trained in this article. The goal of the opponent is to generate a confrontation sample X ^ = σ + X . In this process, a disturbance σ and input sample X are added. If the norm ||·|| describes the differences between the points in the input domain, creating a confrontation sample in model f can be formally turned into the following optimization problem:
X = X + a r g m i n { | ( | φ | ) | : f ( φ + X ) f ( X ) }
There are two reasons why f ( σ + X ) f ( X ) . First, this nonequality allows the classification category error to be obtained, and it makes it possible for large-scale offsets to be performed in the detection box. φ is shifted in the direction of the highest sensitivity of X and can obtain the best disturbance effect on the basis of minimal disturbances. The noise adaptive attack algorithm is used in this article to determine the most suitable disturbance amount. In the early stage of the algorithm, an effective supervision sample needs to be found with the rise in the gradient.
The traditional AdaGrad method uses historical gradients. In this article, an attenuation coefficient is added at the cumulative square gradient stage to control how much of the historically submitted information is obtained. The gradient accumulation is transformed into the moving average of the decaying exponential parameter to optimize the degree of gradient utilization.
Suppose the initial parameter is τ; the training concentration contains a small batch of m samples { x ( 1 ) , , x ( m ) } , and the corresponding target is y ( i ) . The gradient is calculated as:
V t = T τ T g τ 2
η t = α · m t / V t
where α is the initial learning rate, m t = ϕ ( g 1 , g 2 , , g t ) is used to calculate the first-order momentum of the historical gradients, and g 1 represents the first-order gradient.
The gradient optimization method of the adaptive learning rate is adopted here. It makes the parameter learning rate self-adaptive, performs large updates on nonfrequent parameters, and performs smaller updates on frequent parameters. Therefore, it is very suitable for processing sparse data. AdaGrad is more robust than SGD.
It is a second-order gradient summation for all moments, and it is improved later in this paper using a recent second-order gradient sum. The neural network is under nonconvex conditions, so the latter method will perform better in this experiment.

3.4. Attack Noise

The adaptive noise attack algorithm used in this paper adds a large amount of noise to the data, and this noise is normally distributed. The disturbance value Δx is added to the FGSM algorithm to improve the gradient expansion mode. The disturbance value must be as small as possible. The loss function after disturbance value training is larger than that used by the gradient decrease method. At present, the model loss function is set to L. The θ value generated during gradient expansion is determined. It is necessary to find Δx to attack the model and decrease its performance; then, the L value is increased. The gradient expands on X because only one attack sample does not have an obvious impact, so repeated iteration is needed. After completing this process, the model input sample is X ^ , and the iteration leads to f ( X ^ ) f ( X ) .
The expression is as follows:
X t = X t 1 + l r L x t 1
where l r denotes to the learning rate and L / x is the gradient of the loss function. After this step, the attack sample X ^ can be better supervised. The iteration process is optimized with AdaGrad.
Then, η   ~   n   ( μ ,   Δ ) to sample multiple noise types from the normal distribution, where μ = 0. This sampling method guarantees that the input sample x and the original sample x after noise is added satisfy E [ | X X | ] 0 . Then, the performance of Δ x is controlled by controlling the size of hyperparameter δ . Because the normal distribution of the selected sample is 0, the value of the noise to be distributed in the data sample after noise superposition is still stable. Then, ϵ needs to be constrained to ensure that the level of the formal difference in the normal distribution satisfies x x p ϵ . The added noise may not negatively impact the model, and much noise should be sampled and screened. The mathematical expression and its explanation are as follows:
η N ( 0 , δ )
This formula indicates that the N-group variance is a normal distribution of δ averaged to 0. The square difference δ is adjusted according to the desired impact on ϵ . The highest noise η t that can maximize the loss function L ( f ( x + Δ x ) , y ) is selected. Its mathematical expression is:
η t = a r g m a x η H L ( f ( I t 1 + η ) , y t )
where η is the choice of disturbance noise that can maximize the value of L.
I t = H t 1 + φ ( X ( H t 1 + α η t ) )
where I t 1 refers to the attack sample of t 1 , and the algorithm continues attack training. In addition, x ^ ( I t 1 + α η t ) represents the difference in the vector dimensions of the other effective attack samples. The setting of super parameter φ implies the presence of smaller supervision, which can preserve the noise features to the maximum extent. However, this also signals a reduction in the ability to attack, so multiple iterations are needed to find effective attack data. After the selection is completed, the average value of X and the noise need to be determined for our method.
Therefore, this article expands the gradient, and it performs multiple iterations to obtain effective noise to achieve the purpose of the model. In addition, a Gaussian distribution with a mean of 0 is selected to control the level of the noise disturbance. The Gaussian distribution based on the dynamics of ϵ is used to generate candidate interference σ 0 . The samples select the X value with the highest loss function value and supervise the model with that function.
Finally, the minimum noise can be obtained by combining the two steps in the above method.
As shown in Figure 1, with the blue arrow η and after confirming the effective attack sample x ^ , a new attack sample is created. The candidate disturbance is obtained after t secondary iterations.

4. Results

SAR ship data are extremely scarce because they are difficult to collect and interpret. For the public data set used in [30,31,32,33,34], the data scale is far smaller than that of the current popular deep learning data set. There are many data differences: the sizes of the images are different, the polarization methods are different, and the difference in the resolutions is large. In this article, SSDD [35,36,37,38,39,40,41] data sets with complete scenes and many data samples are used as the experimental data sets. In the following experiments, to control the variables and exclude the interference factors that affect the results, we choose to compare the algorithms on the same data set.
We previously designed a statistical adaptation loss function based on attention and SAR ship data enhancement. Based on this model, we conduct related experiments on the confrontation attack model.
Experimental procedure:
To test the generalized performance of the model, the overall data are divided into three parts, i.e., α%, 20% and (100-20-α)% subsets, where α ∈ (20,60). These three subsets are the training data, testing data, and verification data, respectively. In the experiment, to ensure balanced training data and verification data, α = 50.
Moreover, 20% of the data are the regular test data of the model.
The attack samples are selected and added to the α% training data. The mixed α% training data are entered into the model, and the model performs defensive training.
(100-20-α)% of the data are always unchanged, as is the case when verifying the performance of third-party data test models. The reason for this is that it is difficult to find third-party data with similar distributions, sizes and target sizes in the available data sets to use as appropriate verification data. Therefore, using some data in their original format as verification data is a way to adapt to the environment.
The first four sets of data in Figure 2 show that the attack effect directly leads to model misses and misunderstandings. In Figure 2, although the target is still in the enclosure, the detection box has shifted sharply. The detection box shift is also a manifestation of the decline in the accuracy of the detection model after the attack.

5. Data Analysis

5.1. Attack Experiment

Table 1 shows the attack effects of different attack methods.
To evaluate object detection models, the mean average precision (mAP) [41] is used. The mAP of the original data is 97.88 when there is no attack. The attack effect of random noise on the model is not obvious, but the mAP is slightly reduced, by 4.3%.
When our NAA method is used, the checking rate and recall rate are the lowest. Compared with the several other attack methods in the table, our attack method is better. Moreover, the disturbance rate is the lowest, and the results prove that the data after interference are very close to the original image. The success rate of the NAA offense is high, and the generalization performance is the best. Even when the attack power of NAA is not the highest, the performance of the model after the defence procedures are applied is very high.
It can be seen that FGSM is infinitely iterated and damaged. Compared with FGSM, NAA is weak. The success rate of our NAA model is slightly higher than that of several other attack methods. All the attacks cannot be seen with the naked eye. Our purpose is to explore the impact of attack noise on a model when the disturbance rate is very low. The high and low significance of the disturbance rate indicator is not significant, and the naked eye cannot distinguish the images when the disturbance rate is low. This attack method can be used improve the defence performance of detection models so that the overall defence success rate will not decrease significantly.

5.2. Defensive Experiment

Creating simple attack models is not our goal. Our goal is to find a way to resist these attacks. Therefore, a series of defensive experiments are conducted.
The previously mentioned defence experiments are used to show effectiveness of a defence model in combating attack data when they are masked in the training data. The defence effect obtained by the training model is shown in Table 2. Compared with the results in Table 1, it can be seen that the accuracy of the detection model after the defence procedure has improved significantly. The attack samples generated by the NAA method are significant. These samples are better than other attack samples for model defence training.
We are very interested in exploring the performance of the model before and after defence training. To understand the real situation, we performed a set of comparative experiments. Using the data division method mentioned earlier, the overall data are divided into three parts, α%, 20% and (100-20-α)%, where α ( 20 , 60 ) . α% of the data are used as training data, 20% are used as test data, and the remaining (100-20-α)% are used as verification data. In the experiment, to ensure balanced training data and verification data, α = 50 .
Table 3 shows the performance of the model on the 20% test set after using the α% data to fight the attack. This is because of the scarcity of SAR ship data and the very large differences in public data. In general, the variables must be controlled, and the data changes must be objectively reflected. This makes the model perform poorly because the training data are insufficient. However, this does not affect the anti-defence effect of our observation and comparison models. The focus of our attention is the trend of the data change.
The data shown in Table 4 use the α% data training model to verify the detection effect on the (100-20-α)% data set. Attack-response training is not used for the verification data set; it is only used for testing, and the purpose is to compare it with the model after defence training.
For the results shown in Table 5, the α% data are used to train the models, and then the α% data are used as confrontation data for defensive training. The detection effect is verified on the (100-20-α)% data set.
Notably, in Table 3, Table 4 and Table 5, the experimental data designed for Table 4 and Table 5 can be used in multiple tests (cross-verification). In each experiment, the data are verified to ensure the objectivity and fairness of the results.
The experimental results show that compared with the defensive monitoring model, the mAP obtained by attack-sample defence training is improved.
It can be seen from the comparison of Table 4 and Table 5 that NAA confrontation defence training improves the detection performance on the third part of the data. In the comparison experiments, due to the decrease in the amount of training data, the overall data performance is lower than that in Table 1 and Table 2. However, from the overall data, it can still be seen that the generalization ability of the model is increased.

6. Conclusions

Here, we address the original questions of the study. Gaussian distributed noise with a preset average value of 0 is added to the original data as confrontation attack data. Compared with other random noise types, Gaussian distributed noise is more similar to actual noise data. The gradient expansion method is used to combine the noise to fight against attack methods generated by adapting the attack algorithm. This attack method has a strong negative effect on the SAR ship detection model, it can greatly reduce the accuracy of the identification results obtained by the ship model, and it has a lower disturbance rate. The attack data that are screened during training can effectively improve the defence capabilities of the model. The anti-defence ability of defence training and the generalization ability of the model are strong.

Author Contributions

Conceptualization, W.G.; writing—review and editing, W.G.; software, Y.Z.; visualization, Q.L. (Quanyang Liu) and Q.L. (Qi Li); project administration, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Science and Technology Department Project of Jilin Province (under grants No. 20210502021ZP).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Written informed consent has been obtained from the patients to publish this paper.

Data Availability Statement

The SSDD product used in this work is available at: https://github.com/TianwenZhang0825/Official-SSDD.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Oliver, C.J. Understanding Synthetic Aperture Radar Images; Artech House: Boston, MA, USA; London, UK, 1998. [Google Scholar]
  2. Jackson, C.R.; Apel, J.R. Principles of Synthetic Aperture Radar. In Synthetic Aperture Radar: Marine User’s Manual, 1st ed.; Samuel, W., McCandless, A., Jr., Jackson, C.R., Eds.; NOAA: NOAA Central Library: Silver Spring, MD, USA, 2004; Volume 1, pp. 1–24. [Google Scholar]
  3. Intergovernmental Oceanographic Commission. Guide to Satellite Remote Sensing of the Marine Environment [OBSOLETE]; Intergovernmental Oceanographic Commission: Paris, France, 2021. [Google Scholar]
  4. Rani, M.; Masroor, M.; Kumar, P. Remote sensing of Ocean and coastal environment–overview. Remote Sens. Ocean. Coast. Environ. 2021, 2, 1–15. [Google Scholar]
  5. Special Issue “Remote Sensing Techniques in Marine Environment”. Available online: https://www.mdpi.com/journal/jmse/special_issues/wl_remote_sensing (accessed on 10 November 2022).
  6. Bayındır, C.; Frost, J.D.; Barnes, C.F. Assessment and enhancement of sar noncoherent change detection of sea-surface oil spills. IEEE J. Ocean. Eng. 2017, 43, 211–220. [Google Scholar] [CrossRef]
  7. Kavanaugh, M.T.; Bell, T.; Catlett, D.; Cimino, M.A.; Doney, S.C.; Klajbor, W.; Messié, M.; Montes, E.; Muller-Karger, F.E.; Otis, D.; et al. Satellite Remote Sensing and the Marine Biodiversity Observation Network. Oceanography 2021, 34, 62–79. [Google Scholar] [CrossRef]
  8. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  9. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  10. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems 28; Cortes, C., Lawrence, N.D., Lee, D.D., Sugiyama, M., Garnett, R., Eds.; Curran Associates, Inc.: New York, NY, USA, 2015; pp. 91–99. [Google Scholar]
  11. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  12. Samanta, S.; Mehta, S. Towards crafting text adversarial samples. arXiv 2021, arXiv:1707.02812. [Google Scholar]
  13. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  14. Sarkar, S.; Bansal, A.; Mahbub, U.; Chellappa, R. UPSET and ANGRI: Breaking high performance image classifiers. arXiv 2021, arXiv:1707.01159. [Google Scholar]
  15. Chen, P.Y.; Zhang, H.; Sharma, Y.; Yi, J.; Hsieh, C.J. ZOO: Zeroth order optimization based black-box attacks to deep neural networks without training substitute models. In Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security, Dallas, TX, USA, 3 November 2017; ACM: New York, NY, USA, 2017; pp. 15–26. [Google Scholar]
  16. Wang, M.; Wang, H.; Wang, L. Adversarial Examples Generation And Attack On SAR Image Classification. In Proceedings of the 2021 5th International Conference on Innovation in Artificial Intelligence, Xiamen, China, 5–8 March 2021. [Google Scholar]
  17. Kurakin, A.; Goodfellow, I.J.; Bengio, S. Adversarial examples in the physical world. In Artificial Intelligence Safety and Security; Chapman and Hall/CRC: Boca Raton, FL, USA, 2018; pp. 99–112. [Google Scholar]
  18. Dong, Y.; Pang, T.; Su, H.; Zhu, J. Evading defenses to transferable adversarial examples by translation-invariant attacks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4312–4321. [Google Scholar]
  19. Lin, J.; Song, C.; He, K.; Wang, L.; Hopcroft, J.E. Nesterov accelerated gradient and scale invariance for adversarial attacks. arXiv 2019, arXiv:1908.06281. [Google Scholar]
  20. Nesterov, Y. A method for unconstrained convex minimization problem with the rate of convergence o(1/k^2). Sov. Math. Dokl. 1983, 269, 543–547. [Google Scholar]
  21. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar]
  22. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial machine learning at scale. arXiv 2016, arXiv:1611.01236. [Google Scholar]
  23. Tramèr, F.; Kurakin, A.; Papernot, N.; Goodfellow, I.; Boneh, D.; McDaniel, P. Ensemble adversarial training: Attacks and defenses. arXiv 2017, arXiv:1705.07204. [Google Scholar]
  24. Papernot, N.; McDaniel, P.; Wu, X.; Jha, S.; Swami, A. Distillation as a defense to adversarial perturbations against deep neural networks. In Proceedings of the 2016 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2016; pp. 582–597. [Google Scholar]
  25. Jia, X.; Wei, X.; Cao, X.; Foroosh, H. ComDefend: An efficient image compression model to defend adversarial examples. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6077–6085. [Google Scholar]
  26. Liu, Z.; Liu, Q.; Liu, T.; Xu, N.; Lin, X.; Wang, Y.; Wen, W. Feature distillation: DNN-oriented JPEG compression against adversarial examples. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 860–868. [Google Scholar]
  27. Qiu, H.; Zeng, Y.; Zheng, Q.; Guo, S.; Zhang, T.; Li, H. An efficient preprocessing-based approach to mitigate advanced adversarial attacks. IEEE Trans. Comput. 2021. [Google Scholar] [CrossRef]
  28. Moosavi-Dezfooli, S.M.; Fawzi, A.; Frossard, P. Deepfool: A simple and accurate method to fool deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–July 1 2016. [Google Scholar]
  29. Torres, L.; Sant’Anna, S.J.S.; da Costa Freitas, C.; Frery, A.C. Speckle reduction in polarimetric SAR imagery with stochastic distances and nonlocal means. Pattern Recognit. 2014, 47, 141–157. [Google Scholar] [CrossRef] [Green Version]
  30. Huang, L.; Liu, B.; Li, B.; Guo, W.; Yu, W.; Zhang, Z.; Yu, W. OpenSARShip: A dataset dedicated to sentinel-1 ship interpretation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 195–208. [Google Scholar] [CrossRef]
  31. Li, B.; Liu, B.; Huang, L.; Guo, W.; Zhang, Z.; Yu, W. OpenSARShip 2.0: A large-volume dataset for deeper interpretation of ship targets in sentinel-1 imagery. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–14 November 2017; pp. 1–5. [Google Scholar]
  32. Li, J.; Qu, C.; Shao, J. Ship detection in SAR images based on an improved faster R-CNN. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–14 November 2017; pp. 1–6. [Google Scholar]
  33. Wang, Y.; Wang, C.; Zhang, H.; Dong, Y.; Wei, S. A SAR dataset of ship detection for deep learning under complex backgrounds. Remote Sens. 2019, 11, 765. [Google Scholar] [CrossRef] [Green Version]
  34. Wei, S.; Zeng, X.; Qu, Q.; Wang, M.; Su, H.; Shi, J. HRSID: A high-resolution SAR images dataset for ship detection and instance segmentation. IEEE Access 2020, 8, 120234–120254. [Google Scholar] [CrossRef]
  35. Zhang, T.; Zhang, X.; Li, J.; Xu, X.; Wang, B.; Zhan, X.; Xu, Y.; Ke, X.; Zeng, T.; Su, H.; et al. SAR Ship Detection Dataset (SSDD): Official release and comprehensive data analysis. Remote Sens. 2021, 13, 3690. [Google Scholar] [CrossRef]
  36. Zhang, T.; Zhang, X.; Ke, X.; Zhan, X.; Shi, J.; Wei, S.; Pan, D.; Li, J.; Su, H.; Zhou, Y.; et al. LS-SSDD-v1.0: A Deep Learning Dataset Dedicated to Small Ship Detection from Large-Scale Sentinel-1 SAR Images. Remote Sens. 2020, 12, 2997. [Google Scholar] [CrossRef]
  37. Zhang, T.; Zhang, X.; Shi, J.; Wei, S. HyperLi-Net: A hyper-light deep learning network for high-accurate and high-speed ship detection from synthetic aperture radar imagery. ISPRS J. Photogramm. Remote Sens. 2020, 167, 123–153. [Google Scholar] [CrossRef]
  38. Zhang, T.; Zhang, X. ShipDeNet-20: An only 20 convolution layers and< 1-MB lightweight SAR ship detector. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1234–1238. [Google Scholar] [CrossRef]
  39. Zhang, T.; Zhang, X. High-speed ship detection in SAR images based on a grid convolutional neural network. Remote Sens. 2019, 11, 1206. [Google Scholar] [CrossRef] [Green Version]
  40. Zhang, T.; Shi, J.; Wei, S. Depthwise Separable Convolution Neural Network for High-Speed SAR Ship Detection. Remote Sens. 2019, 11, 2483. [Google Scholar] [CrossRef] [Green Version]
  41. Special Issue “Evaluating Object Detection Models Using Mean Average Precision (mAP)”. Available online: https://blog.paperspace.com/mean-average-precision/ (accessed on 20 January 2023).
Figure 1. The η attack generation process.
Figure 1. The η attack generation process.
Sensors 23 02266 g001
Figure 2. Comparison of the detection effect before and after the attack (The (a,c,e,g,i,k,m) panels show the detection effects of the models before the attacks. The (b,d,f,h,j,l,n) panels show the detection effects of the models after the attacks.).
Figure 2. Comparison of the detection effect before and after the attack (The (a,c,e,g,i,k,m) panels show the detection effects of the models before the attacks. The (b,d,f,h,j,l,n) panels show the detection effects of the models after the attacks.).
Sensors 23 02266 g002aSensors 23 02266 g002b
Table 1. Comparison of the effects of different attack methods.
Table 1. Comparison of the effects of different attack methods.
AttackPrecisionRecallSuccess RatemAP
Original97.0297.62097.88
Random Noise93.1293.366.293.58
FGSM18.6519.1279.2319.21
AdvGAN16.5317.1280.1817.32
SI-NI15.5215.6882.6916.12
TIM14.3214.6584.7614.87
NAA11.2711.8382.511.86
Table 2. Defence effect comparison of different attack methods.
Table 2. Defence effect comparison of different attack methods.
AttackPrecisionRecallmAP
FGSM95.2895.7295.94
SI-NI95.8796.2896.42
NAA96.2196.7196.82
Table 3. Test of the detection effect on the test sets when no defence training is applied.
Table 3. Test of the detection effect on the test sets when no defence training is applied.
AttackPrecisionRecallmAP
FGSM12.1212.2312.36
SI-NI10.2110.5210.72
NAA5.695.956.42
Table 4. Verification data test results when no defence training is applied.
Table 4. Verification data test results when no defence training is applied.
AttackPrecisionRecallmAP
Original78.6278.8279.02
Table 5. Defence training verification test results.
Table 5. Defence training verification test results.
AttackPrecisionRecallmAP
FGSM78.6378.8379.03
SI-NI79.1779.2479.23
NAA83.5283.3684.58
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, W.; Liu, Y.; Zeng, Y.; Liu, Q.; Li, Q. SAR Image Ship Target Detection Adversarial Attack and Defence Generalization Research. Sensors 2023, 23, 2266. https://doi.org/10.3390/s23042266

AMA Style

Gao W, Liu Y, Zeng Y, Liu Q, Li Q. SAR Image Ship Target Detection Adversarial Attack and Defence Generalization Research. Sensors. 2023; 23(4):2266. https://doi.org/10.3390/s23042266

Chicago/Turabian Style

Gao, Wei, Yunqing Liu, Yi Zeng, Quanyang Liu, and Qi Li. 2023. "SAR Image Ship Target Detection Adversarial Attack and Defence Generalization Research" Sensors 23, no. 4: 2266. https://doi.org/10.3390/s23042266

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop