Next Article in Journal
Applications of Clustering Methods for Different Aspects of Electric Vehicles
Next Article in Special Issue
Underwater Noise Modeling and Its Application in Noise Classification with Small-Sized Samples
Previous Article in Journal
Infrared and Visible Image Homography Estimation Using Multiscale Generative Adversarial Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Mode Channel Position Attention Fusion Side-Scan Sonar Transfer Recognition

1
Acoustic Science and Technology Laboratory, Harbin Engineering University, Harbin 150001, China
2
College of Underwater Acoustic Engineering, Harbin Engineering University, Harbin 150001, China
3
Key Laboratory of Marine Information Acquisition and Security, Harbin Engineering University, Ministry of Industry and Information Technology, Harbin 150001, China
4
College of Internet of Things Engineering, Hohai University, Changzhou 213022, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(4), 791; https://doi.org/10.3390/electronics12040791
Submission received: 12 January 2023 / Revised: 1 February 2023 / Accepted: 2 February 2023 / Published: 4 February 2023

Abstract

:
Side-scan sonar (SSS) target recognition is an important part of building an underwater detection system and ensuring a high-precision perception of underwater information. In this paper, a novel multi-channel multi-location attention mechanism is proposed for a multi-modal phased transfer side-scan sonar target recognition model. Optical images from the ImageNet database, synthetic aperture radar (SAR) images and SSS images are used as the training datasets. The backbone network for feature extraction is transferred and learned by a staged transfer learning method. The head network used to predict the type of target extracts the attention features of SSS through a multi-channel and multi-position attention mechanism, and subsequently performs target recognition. The proposed model is tested on the SSS test dataset and evaluated using several metrics, and compared with different recognition algorithms as well. The results show that the model has better recognition accuracy and robustness for SSS targets.

1. Introduction

Side-scan sonar (SSS) generates pseudo-color images of varying intensities by recording the intensity of backscattered sound waves from the ocean floor. It has the characteristics of wide coverage and high resolution and can not only map seabed topography but also image underwater targets (wrecks, aircraft). It is the main sensor used in autonomous underwater vehicles (AUV). In recent years, automatic target recognition (ATR) based on SSS data is used to quickly and automatically classify underwater targets [1]. Real-time detection and identification of submarine objects using low-cost and low-risk AUVs is an issue of great interest to Navies around the world [2,3]. However, the signal-to-noise ratio of sonar images is extremely low, and the target information is seriously polluted due to the existence of ocean noise, reverberation noise and speckle noise [4,5,6,7]. In addition, with the increase of the operating frequency of SSS, the absorption of sound energy by seawater increases. Consequently, the sound absorption loss of sound waves in seawater is large and the sonar images exhibit the low target contrast phenomenon [8,9], which directly affects the target recognition performance.
The deep learning (DL) technology can convolve the input image through multiple convolution kernels of different scales, filtering the image and realizing the feature extraction of the input image. The focus of a deep learning object recognition system is to design the structure of the convolutional neural network. The input image is extracted from shallow to deep through the designed network structure, and the target is identified by the classifier according to the features extracted by the final layer. It can be observed that for the target recognition method based on the convolutional neural network (CNN) [10,11,12,13], the most important factor is to extract stable features. The instability of features can be reflected in the problems of insufficient data samples and features susceptible to noise interference. If the designed network is huge and the network parameters have high magnitudes, the recognition network will have an overfitting problem [14]. This means that the recognition performance of the network on the training and test sets will be very different, and the features extracted under noise interference will have noisy features. Such features have a negative impact on final recognition. Therefore, in order to obtain a good SSS target recognition performance, it is necessary to mitigate the problem associated with insufficient training samples and study the generation of stable features in the network design process.

1.1. Literature Review

Recent research on SSS target recognition has shown that the target recognition method based on CNNs has outperformed traditional machine learning methods, including the fuzzy logic method, K nearest neighbor, support vector machines, etc., [15,16,17]. With an increasing scale, the recognition network can extract deeper features from images to obtain richer feature information. Therefore, a research direction for target recognition is to design a deeper recognition network model. Simonyan et al., Alex et al., He et al. and Xu et al. proposed VGG [18], AlexNet [19], ResNet [20], and DesNet [21] methods, respectively. On the basis, the ResNet152 and Desnet161 methods were proposed to increase the number of network layers and improve the ability of feature extraction [22,23]. However, the aforementioned methods require a large number of data samples in the process of parameter learning. As it is difficult to obtain sample data for SSS data, the lack of training samples limits the application of the methods proposed in [20,21,22,23] for SSS target recognition. In contrast, sample transfer learning (STL) adopts the method of source data set to train the main network body of the model, and target data set to fine-tune the head network, which can effectively solve the problem of insufficient training samples. This research direction has received a significant amount of attention recently. For example, Ye et al. applied the VGG-11 and ResNet-18 models to the underwater target identification of SSS images and adopted optical data pre-training and SSS data fine-tuning to improve the low recognition rate caused by insufficient data samples [24]. Huo et al. used semi-synthetic and SSS data to fine-tune the VGG parameters in order to further improve the sample identification accuracy [25]. Tang et al. detected shipwreck targets using SSS based on the trained DarkNet-53 network [26]. Fuchs et al. applied the transfer recognition method to the target recognition of forward-looking sonar (FLS) [27]. Zhang et al. adopted the YOLOv5 model to detect FLS targets [28].
All the above studies have proved the effectiveness of STL in overcoming the problem of insufficient samples. These studies use different data sets and target data sets to pre-train different backbone networks and fine-tune the head network according to different task requirements, respectively, and finally achieve detection and recognition. However, the SSS images have strong noise interference that is considerably different from the optical images. This difference in distribution causes the features learned by the pre-trained network to significantly limit the recognition accuracy of the head network. In order to overcome this shortcoming, we should study the use of different modal data sets to train the network in different stages, so as to alleviate the problem of insufficient recognition accuracy caused by data differences.
Zhen et al. used synthetic aperture radar (SAR) data to train the middle layer of the network. At the same time, the attention mechanism was utilized to make the features extracted by the network more representative, which improved the anti-noise performance and network recognition accuracy [29]. Although this method was effective, the attention weights of different positions of the same channel were the same in the attention network design. Furthermore, the attention weights of different channels of the same position were also the same. This network design reduced the attention diversity in the network. To address the aforementioned limitations, it is necessary to carry out research on multi-channel multi-location attention networks.

1.2. Novel Contributions

Aiming at the above research gaps, a new multi-channel multi-location attention model (MMA) is proposed for SSS target recognition. In this model, the ResNet network is used as the backbone network, and the ImageNet dataset and the SAR dataset are used to train different stages of the backbone network to improve its data adaptability to SSS. Finally, the MMA model and head recognition network are trained using the SSS dataset to produce accurate and robust recognition results. The novel scientific contributions of this study are as follows:
1. A deep learning model for SSS object recognition is proposed. The model structure can learn the backbone network parameters in stages through a variety of data sets, reduce the distribution difference between the ImageNet data and SSS data, and improve the recognition accuracy for SSS targets. The effectiveness of the proposed recognition model is verified on a measured SSS target dataset (https://toscode.gitee.com/wangjian19870118/ssd-dataset.git, accessed on 27 January 2023).
2. In the previous research on transfer target recognition, only the fully connected layer of the head network was transferred and adjusted, which limited the learning scope. There was no key feature extraction of the key positions and key channels of the SSS target. Unlike previous research, a multi-channel and multi-position attention mechanism is proposed. During the process of extracting SSS target features, different channel attention factors can be set for channels at different positions, and positions can be set at different channel attention factors. These diverse attention factors can achieve all-round acquisition of SSS target features.
3. An integration strategy based on multimodal staged transfer is proposed to enhance the generalization performance of the model. Unlike the existing research, this strategy can enable the backbone network to simultaneously have better classification accuracy for multiple different modal data, so that the SSS target recognition has better recognition performance.

2. Methods

The multi-modal multi-channel multi-position attention transfer recognition model (3MATR) proposed in this paper consists of two stages: (1) multi-modal staged transfer stage and, (2) SSS image attention recognition stage. Figure 1 shows the specific framework of the proposed model.
1.
Stage-I: multi-modal staged transfer.
The training data set is divided into an optical ImageNet data set and the SAR data set. The latter is used to train the front part of the backbone network (ResNet152), and the former is used to train the middle part of the network. This stage enables the backbone network to accurately identify noisy datasets.
2.
Stage-II: multi-channel multi-location attention model target recognition.
The backbone network learned from multiple datasets is combined with the attention mechanism network (AMN) and the recognition network to build an integrated learning model. The SSS data are used to learn the parameters of the network at this stage. During the parameter learning process, attention is allocated to different positions of different channels to obtain the important features of the target. Finally, the SSS target is accurately identified.

2.1. Multi-Modal Staged Transfer

The image size of the training data set is 224 × 224, and the backbone network adopts the ResNet series network structure. Figure 2 shows the local structure. The constant mapping in the ResNet structure can solve the vanishing gradient problem and model degradation during the parameter learning process. In this study, the SAR and ImageNet datasets are used to train the front and middle sections of ResNet152, respectively. The ResNet152 network consists of multiple 3 × 3 and 1 × 1 convolution kernels. The former kernel unit extracts the features of the image information, and the latter kernel can upgrade and reduce the dimensionality of the feature channel through the integration of cross-channel features. Figure 3 shows the ResNet152 network structure diagram.
The ResNet152 model is widely used in target recognition problems. The backbone network is composed of five parts: C o n v 1 , C o n v 2 _ x ( x = 1 3 ) , C o n v 3 _ x ( x = 1 8 ) , C o n v 4 _ x ( x = 1 36 ) , and C o n v 5 _ x ( x = 1 3 ) , where C o n v 1 and C o n v 2 _ x ( x = 1 3 ) form the front segment of the network using SAR data for parameter learning, and C o n v 3 _ x , and C o n v 4 _ x form the middle segment of the network using ImageNet data for parameter learning. C o n v 5 _ x ( x = 1 3 ) and the remaining parts of the network use SSS data for transfer parameter learning. In this way, the purpose of phased transfer learning is achieved. Table 1 shows the details of the convolution kernel settings in ResNet152.

2.2. Multi-Channel Multi-Location Attention Model

The convolution operation in the network uses convolution kernels of different sizes to extract features of local areas, therefore, it cannot perceive the entire position space. A larger position space can be perceived by using a deeper network structure. Although this method perceives the entire location space, it splits the spatial information and can easily lead to a loss of feature information. In this paper, we propose a method that can collect global features and preserve positional relationships. This method can help subsequent convolutional layers perceive the entire space and capture complex feature relationships. At the same time, we propose a multi-channel and multi-location attention mechanism, which can fully consider the channel attention factors of different positions and the position attention factors of different channels. In this way, compact key features can be obtained from complex global images, and the importance and robustness of the proposed features can be enhanced. Next, we describe the architecture and theory of the proposed method in detail.

2.2.1. Channel Attention Mechanism Considering Location (CAMCL)

The design of the CAMCL attention model proposed in this paper is based on the QKV model [30,31,32,33] in the self-attention mechanism. Let the input image be X R c × h × w , denoting the input tensor of a spatiotemporal (2D) convolutional layer. The number of channels is represented by c, and h and w represent the height and width of the spatial scale, respectively. The feature map of X through the convolution kernel is denoted by A , B , V , where A , B , V are three different layers represented as follows:
A = ϕ X ; W ϕ B = θ X ; W θ V = φ X ; W φ
where W ϕ , W θ and W φ are the corresponding network parameters. The layers A , B and V are expanded in the spatial dimension, where B c × h w = [ b 1 , b 2 , b h w ] and V c × h w = [ v 1 , v 2 , v h w ] . In order to calculate the channel attention factor, this paper uses the matrix product A B T to calculate the channel correlation matrix G c × c , which is described as
G c × c = g 1 , g 2 , g c = A B T = a 1 , a 2 , a h w b 1 , b 2 , b h w T = a 1 , a 2 , a h w b 1 T b 2 T b h w T = a 1 ¯ a 2 ¯ a c ¯ b 1 ¯ , b 2 ¯ , b c ¯ = a 1 ¯ b 1 ¯ a 1 ¯ b 2 ¯ a 1 ¯ b c ¯ a 2 ¯ b 1 ¯ a 2 ¯ b 2 ¯ a 2 ¯ b c ¯ a c ¯ b 1 ¯ a c ¯ b 2 ¯ a c ¯ b c ¯ a i ¯ 1 × h w b j ¯ h w × 1 i , j = 1 , 2 , c
a i ¯ b j ¯
where a i ¯ b j ¯ indicates the attention factor between channels i and j. The image is processed by channel attention, which can be expressed as Z c × h w = G c × c V c × h w . The image after channel attention processing is obtained by reshaping Z in order to take into account the position factor while performing channel attention. In Equation (2), b j ¯ ( j = 1 , , h w ) is transformed into an effective position attention weight vector using s o f t m a x ( ) . At the same time, V c × h w = [ v 1 ¯ ; v 2 ¯ , , v c ¯ ] also performs the transformation identical to A = a 1 ¯ ; a 2 ¯ , a c ¯ . Therefore, using Equation (2), the channel attention mechanism considering the location factor can be expressed as follows:
Z ˜ c × h w = G ˜ c × c V ˜ c × h w = a 1 ¯ s o f t max ( b 1 ¯ ) a 1 ¯ s o f t max ( b 2 ¯ ) a 1 ¯ s o f t max ( b c ¯ ) a 2 ¯ s o f t max ( b 1 ¯ ) a 2 ¯ s o f t max ( b 2 ¯ ) a 2 ¯ s o f t max ( b c ¯ ) a c ¯ s o f t max ( b 1 ¯ ) a c ¯ s o f t max ( b 2 ¯ ) a c ¯ s o f t max ( b c ¯ ) s o f t max ( v 1 ¯ ) s o f t max ( v 2 ¯ ) s o f t max ( v c ¯ ) a i ¯ 1 × h w b j ¯ h w × 1 v k ¯ 1 × h w i , j , k = 1 , 2 , c
Figure 4 shows the calculation block diagram of the channel attention mechanism considering the positions.

2.2.2. Location Attention Mechanism Considering Channel (LAMCC)

In order to calculate the location attention factor, this paper uses the matrix multiplication A T B to calculate the location correlation matrix F h w × h w = [ f 1 , f 2 , , f h w ] as follows:
F h w × h w = f 1 , f 2 , , f h w = A T B = a 1 , a 2 , , a h w T b 1 , b 2 , , b h w = a 1 T a 2 T a h w T b 1 , b 2 , , b h w = a 1 T b 1 a 1 T b h w a 1 T b h w a 2 T b 1 a 2 T b h w a 2 T b h w a h w b 1 a h w b h w a h w b h w a i R c × 1 b j R c × 1 i , j = 1 , 2 , h w
where a i T b j indicates the attention factor between locations i and j. The location attention operation of an image can be expressed as K h w × c = F h w × h w ( V c × h w ) T , which can be reshaped to obtain the image after location attention processing. In order to consider the channel attention factor while performing position attention, we adopt an idea similar to CAMCL, and express Equation (4) as follows using the s o f t m a x ( ) function:
K ˜ h w × c = F ˜ h w × h w V ˜ c × h w T = a 1 T s o f t m a x b 1 a 1 T s o f t m a x b 2 a 1 T s o f t m a x b h w a 2 T s o f t m a x b 1 a 2 T s o f t m a x b 2 a 2 T s o f t m a x b h w a h w T s o f t m a x b 1 a h w T s o f t m a x b 2 a h w T s o f t m a x b h w s o f t m a x v 1 T s o f t m a x v 2 T s o f t m a x v h w T a i R c × 1 b j R c × 1 v k R c × 1 i , j , k = 1 , 2 , h w
Figure 5 shows the calculation block diagram of the location attention mechanism considering the channel.

2.3. Data Description

2.3.1. SSS Dataset

A target image dataset collected via imaging sonar is used to conduct the experiments. The dataset used in the experiment includes three types of image targets as follows: 66 planes, 484 ships, and 578 other pictures. Figure 6 shows some example dataset images. It can be observed that the image has strong noise interference, which causes considerable difficulty in accurate target identification.

2.3.2. Experimental Dataset Preprocessing

The data are randomly extracted for creating training and test datasets, where 70% of the data are used as training data and the remaining 30% are used as test data. Table 2 shows the specific data allocation for each category. In order to reduce the influence of random initialization of parameters on the recognition performance, values obtained over repeated experiments are averaged and used for the final evaluation of the recognition performance. The data types in the dataset are unbalanced, with the largest being the 578 samples of other target types and only 66 plane samples. The unbalanced data will cause the classifier to favor categories with large sample sizes, and the recognition rates of the small sample categories will be poor. At the same time, the total number of samples in the data set is 1128, and the network trained using this dataset will exhibit overfitting.
In this experiment, in order to reduce the impact of unbalanced data and overfitting on the evaluated algorithms, some basic data augmentation methods (flip, rotation, cropping) are also applied to preprocess the data. These include center crop, left bottom crop, left top crop, right bottom crop, right top crop, equal height stretch, equal width stretch, contrast transformation, rotation (45 , 90 , 135 , 180 , 225 , 270 , 315 ), left and right flip. Figure 7 shows the image transformation result and Figure 8 shows the image data used in the experiment. It can be observed from the figures that different types of images have different appearances and distributions.

2.4. Evaluation Metric

We use accuracy to evaluate the criterion for evaluating the model performance. The training network in this paper uses the ResNet152 network. The network uses SAR images and optical images to train the front section and back end of the backbone network, respectively, and uses SSS images to train the head network. The batch size is 30, the learning rate is 0.001, and model validation is performed 10 times using the cross-validation method, The criterion for assessing model performance is the average overall accuracy (OA), which is the percentage of all correct positive classifications and represents the overall classification performance. The OA is calculated as in Equation (6).
OA = i C N i i N
where N i i is the number of test samples that should have been classified as class i and are classified as class i in the actual classification results, c refers to the categories of labels in the test samples, and N is the total number of test samples.

2.5. Experimental Platform

Experiments are run on a Microsoft Windows 10 operating system with an NVIDIA GTX TITAN-XP GPU and 64 GB of memory. Python 3.6.8 version is used to design the network structure.

3. Results and Discussion

In this part, the robustness and effectiveness of the proposed method are verified through comparative experiments and analysis. The method is compared with traditional DL recognition methods to verify the improvement effect of the algorithm on the recognition rate. In addition, this algorithm is compared with related transfer learning (TL) algorithms.
The OA index of the test data set is used to compare the state-of-the-art (SOTA) methods with the method proposed in this paper and analyze the related performance of the results. At the same time, data ablation experiments, CAMCL and LAMCC ablation experiments are used to analyze the performance of the network structure, and the method proposed in this paper is verified using the forward-looking sonar (FLS) data set for sample diversity.
(1)
Comparison of our method with traditional DL models
DenseNet, ResNet, VGGNet and other series methods are the most classical and widely used network structures for target recognition. Therefore, this paper chooses these methods for comparative analysis and lists their details and performance. Experimental results are shown in Figure 9 and Figure 10.
As Table 3 shows, the model structures of VGGNet16 and VGGNet19 are simple and easy to train; however, their actual recognition performance is not ideal, and their recognition accuracy is only 92.56%. As shown in Figure 11, during the training process of VGGNet16 and VGGNet19 models, the recognition accuracy curves exhibit instability and jitter, as shown in Figure 11a,b. In particular, the VGGNet19 network shows a downward trend with an increasing number of epochs, and the final recognition rate is only 60%, as shown in Figure 11b. DenseNet (201, 121, 169) and ResNet (152, 101, 50) network models have similar recognition accuracies, and the ResNet101 has the optimal recognition accuracy, which is equal to 96.41%. At the same time, the recognition accuracy of the ResNet network layer decreases slightly from 96.41% for 101 layers to 95.13% for 152 layers, and the recognition accuracy of the DenseNet network layer decreases from 96.15% for 169 layers to 95.02% for 201 layers. This is because as the number of network layers increases, the features extracted by the network tend to be more abstract and detailed. As the SSS image has a large noise interference, the extracted detailed features include noise features, which directly affect the recognition performance of the network.
(2)
Comparison of methods for the classification of SSS images
A few SOTA methods for SSS target recognition have been proposed in [34,35,36]. These methods use different data enhancement methods to increase the number of training data samples through semi-synthetic data [36] and improve the learning and recognition capabilities of the network, which increases the recognition accuracy significantly. Table 4 shows that using synthetic data and VGG19Net fine-tuning network method, the highest target recognition accuracy of 97.76% can be achieved. The recognition accuracy of other methods is reduced to varying degrees due to the insufficient number of network layers.
As the time consumed by the synthetic data increases according to the complexity of the synthetic model, the time consumed for processing the synthetic data is relatively large. It takes a considerable amount of time to prepare the data before training the recognition network. In underwater recognition tasks that do not require a long training time, it is not appropriate to use synthetic data to train a network for tasks with low time complexity, because the time consumed by the synthetic data will be calculated as a part of the recognition time. The staged attention transfer recognition method proposed in this paper can significantly improve the target recognition accuracy rate to 98.72% through the target domain data attention enhancement method without any time loss caused by additional sample generation. This recognition performance is the best compared to other methods, showing an increase of 1%.
(3)
Comparison of different backbones for the classification of SSS image.
The ablation experiment method is used to pre-train a variety of backbone networks using ImageNet data, and the transfer recognition method is used to fine-tune the parameters of the head network using SSS data to analyze the recognition performance. In order to fully analyze the positive effect of transfer recognition on SSS target recognition, we supplemented AlexNet, GoogleNet and other shallow network transfer recognition methods based on the above traditional network model (ResNet, DenseNet, VGGNet) for comparative analysis. Figure 12, Figure 13 and Figure 14 show the recognition rate and loss curves of the partial recognition network, and Table 5 shows the optimal recognition rate results with different backbone network transfer recognition methods.
Table 5 shows that different backbone networks used to identify SSS targets produce different results: VGGNet16, ResNet152, and Resnet101 have a recognition accuracy rate higher than 98%, while ResNet18 and DenseNet169 have relatively low recognition accurate rates of only 91.86% and 91.54%, respectively. This is because the number of network layers is small for ResNet18: there are only 18 layers that are not sufficient for feature extraction. For DenseNet121, DenseNet169 and DenseNet201, the values of OA are 96.41%, 91.54% and 89.23%, respectively. It can be concluded that as the number of network layers increases, the recognition accuracy rate shows a decreasing trend. This phenomenon is caused by a difference between the training data of the backbone network and the SSS data. As the number of network layers increases, the network extracts more detailed and abstract features. The SSS data have a significant amount of noise interference. This noise exists in the details of the image. The deeper network extracts the noisy features during the extraction of the detailed features. These noisy features have a negative impact on the target recognition performance. Therefore, as the network structure becomes more complex and the number of network layers increases, the recognition rate tends to decline.
The method proposed in this paper comprehensively considers the influence of the number of network layers and network structure on the recognition performance and uses ResNet152 as the backbone network to learn the network parameters and identify the SSS targets. It obtains the best recognition accuracy rate of 98.72%. At the same time, Figure 12, Figure 13 and Figure 14 show that the recognition rate curves of the existing methods have a large variance, whereas the variance of the recognition rate curve of the proposed method is relatively small. This behavior indicates that the recognition performance of the method proposed in this paper is more stable.
(4)
Ablation experiments for backbone network trained on different data sets.
In order to verify the DL strategy problem, this paper uses SAR, ImageNet, and SSS data to conduct ablation experiments on the sub-network structure of the backbone network. Table 6 shows the experimental results.
Table 6 shows that the recognition performance of the backbone networks trained only with ImageNet and SAR data set is poor, achieving OA values of only 58.97% and 77.95%. The recognition performance of the backbone network trained with two data sets is better, and the OA of the training strategy (ImageNet+SAR) is 1.8% lower than that proposed in this paper (SAR+ImageNet).
(5)
CAMCL and LAMCC ablation experiments.
In order to verify the performance of the CAMCL and LAMCC attention fusion methods, this paper conducts attention mechanism ablation experiments. Experiments were carried out on three methods: only CAMCL, only LAMCC, and CAMCL+ LAMCC with ResNet152 as the backbone network.
Table 7 shows that the optimal OA values for only CAMCL and only LAMCC are 97.69% and 97.18%, respectively. The optimal OA value for CAMCL+ LAMCC is 98.72%, signifying an increase in the recognition rates by 1.03% and 1.54% respectively. Therefore, the use of CAMCL and LAMCC attention fusion mechanisms is helpful to improve recognition performance.
(6)
Application of FLS Target Recognition.
The method proposed in this paper can not only recognize SSS targets but can also be applied to FLS target recognition. In order to verify the effectiveness of the proposed method with respect to a higher number of target recognition categories, this paper uses Forward Looking Sonar (FLS) data for experiments. Among them, the total number of samples is 3192, the number of target categories is 10, the number of training sets is 2231, and the number of test sets is 961. Figure 15 shows the data sample diagram and Table 8 shows the experimental results.
Table 8 shows that the recognition rates of the DenseNet201, DenseNet121 and DenseNet169 are equal to 89.07%, 88.87% and 89.91%, respectively, which are slightly higher than those obtained using the ResNet series algorithms. The recognition rates of DenseNet and ResNet algorithms are above 88%. VGGNet16 achieves the best OA = 90.63% among the reference algorithms, and VGGNet19 performs the worst with OA = 85.22%. The method proposed in this paper is the best among all the comparison algorithms, which shows its good FLS target recognition performance.

4. Conclusions and Future Work

An important part of the sonar system that provides accurate target perception information is the SSS underwater target recognition. In this study, a new 3MATR model was provided for SSS target recognition, using the ResNet152 network as the backbone network of the model. The strategy of phased transfer, ImageNet and SAR data were collectively utilized to learn the parameters of the backbone network. In addition, the CAMCL and LAMCC attention mechanisms were used to perform the final and accurate recognition of SSS data in the head network. The proposed model was analyzed from different aspects and its performance was compared with many existing identification algorithms. The following conclusions were obtained:
1. The proposed multi-modal multi-channel multi-position attention mechanism transfer recognition method effectively improved the SSS target recognition accuracy. Its basic principle was to train different stages of the network through ImageNet, SAR, and SSS data sets to enhance the anti-noise performance of the network against noisy backgrounds, and simultaneously add an attention mechanism to extract key features in the SSS training stage.
2. The 3MATR model provided advantages in SSS target recognition, which resulted in significantly better recognition rates than other backbone networks. Its transfer recognition performance was higher than other transfer learning target recognition methods.
The limitation of this method was the difference in data distributions for different modalities in the transfer model. In future work, we envisage overcoming the differences in data distribution by improving the model and further improving the recognition performance.

Author Contributions

Conceptualization, J.W. and H.L.; Methodology, J.W. and G.H.; Software, J.W.; Validation, J.W. and H.L.; Formal analysis, J.W. and C.L.; Investigation, J.W. and Y.W.; Resources, H.L.; Data curation, H.L.; Writing—original draft preparation, J.W.; Writing—review and editing, J.W. and H.L.; Visualization, J.W.; Supervision, G.H. and H.L.; Project administration, G.H. and H.L.; Funding acquisition, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China under grant nos. U1906218 and 2021YFC3101803; in part by Natural Science Foundation of Heilongjiang Province, under grant no. ZD2020D001; and in part by key areas of research and development plan key projects of Guangdong Province under grant no. 2020B1111010002.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Access to the data will be available at https://toscode.gitee.com/wangjian19870118/ssd-dataset.git (accessed on 16 November 2021).

Acknowledgments

We would like to thank Guanying Huo, L-3 Klein Associates, EdgeTech, Lcocean, Hydro-tech Marine, and Tritech for their great support for providing the valuable real side-scan sonar images.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sawas, J. Automatic Target Recognition in Sonar Imagery Using a Cascade of Boosted Classifiers. Ph.D. Thesis, Heriot-Watt University, Edinburgh, UK, 2015. [Google Scholar]
  2. Isaacs, J.C. Sonar automatic target recognition for underwater UXO remediation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 134–140. [Google Scholar]
  3. Stack, J. Automation for underwater mine recognition: Current trends and future strategy. In Proceedings of the Detection and Sensing of Mines, Explosive Objects, and Obscured Targets XVI, Orlando, FL, USA, 25–29 April 2011; Volume 8017, pp. 205–225. [Google Scholar]
  4. Bruna, M.A.; Pate, D.J.; Cook, D. Synthetic aperture sonar speckle noise reduction performance evaluation. J. Acoust. Soc. Am. 2018, 143, 1856. [Google Scholar] [CrossRef]
  5. Chen, W.; Gu, K.; Lin, W.; Xia, Z.; Le Callet, P.; Cheng, E. Reference-free quality assessment of sonar images via contour degradation measurement. IEEE Trans. Image Process. 2019, 28, 5336–5351. [Google Scholar] [CrossRef] [PubMed]
  6. Huo, G.; Yang, S.X.; Li, Q.; Zhou, Y. A robust and fast method for sidescan sonar image segmentation using nonlocal despeckling and active contour model. IEEE Trans. Cybern. 2016, 47, 855–872. [Google Scholar] [CrossRef]
  7. Guo, H.; Lei, X.; Zhao, H.; Jiao, S.; Hou, Y. A morphological filter for despeckling of a sonar image. Chin. J. Sci. Instrum. 2015, 36, 654–660. [Google Scholar]
  8. Pinto, M.A. Split-beam Range-gated Doppler Velocity Sonar for Operations at High Altitude above the Seabed. In Proceedings of the OCEANS 2018 MTS/IEEE Charleston, Charleston, SC, USA, 22–25 October 2018; pp. 1–6. [Google Scholar]
  9. Pinto, M.A.; Verrier, L. Interferometric Doppler Velocity Sonar for Low Bias Long Range Estimation of Speed Over Seabed. IEEE J. Ocean. Eng. 2022, 47, 767–779. [Google Scholar] [CrossRef]
  10. Wang, X.; Jiao, J.; Yin, J.; Zhao, W.; Han, X.; Sun, B. Underwater sonar image classification using adaptive weights convolutional neural network. Appl. Acoust. 2019, 146, 145–154. [Google Scholar] [CrossRef]
  11. Yan, J.; Meng, J.; Zhao, J. Real-time bottom tracking using side scan sonar data through one-dimensional convolutional neural networks. Remote Sens. 2019, 12, 37. [Google Scholar] [CrossRef]
  12. Yu, F.; He, B.; Li, K.; Yan, T.; Shen, Y.; Wang, Q.; Wu, M. Side-scan sonar images segmentation for AUV with recurrent residual convolutional neural network module and self-guidance module. Appl. Ocean. Res. 2021, 113, 102608. [Google Scholar] [CrossRef]
  13. Jin, L.; Liang, H.; Yang, C. Accurate underwater ATR in forward-looking sonar imagery using deep convolutional neural networks. IEEE Access 2019, 7, 125522–125531. [Google Scholar] [CrossRef]
  14. Xu, Q.; Zhang, M.; Gu, Z.; Pan, G. Overfitting remedy by sparsifying regularization on fully-connected layers of CNNs. Neurocomputing 2019, 328, 69–74. [Google Scholar] [CrossRef]
  15. Chaudhary, A.; Tiwari, V.; Kumar, A. A novel intrusion detection system for ad hoc flooding attack using fuzzy logic in mobile ad hoc networks. In Proceedings of the International Conference on Recent Advances and Innovations in Engineering (ICRAIE-2014), Jaipur, India, 9–11 May 2014; pp. 1–4. [Google Scholar]
  16. Peeples, J.; Cook, M.; Suen, D.; Zare, A.; Keller, J. Comparison of possibilistic fuzzy local information c-means and possibilistic k-nearest neighbors for synthetic aperture sonar image segmentation. In Proceedings of the Detection and Sensing of Mines, Explosive Objects, and Obscured Targets XXIV, Baltimore, MD, USA, 15–17 April 2019; Volume 11012, pp. 211–220. [Google Scholar]
  17. Park, J.; Hwang, C.; Bae, K. Analysis of target classification performances of active sonar returns depending on parameter values of SVM kernel functions. J. Korea Inst. Inf. Commun. Eng. 2013, 17, 1083–1088. [Google Scholar]
  18. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  19. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  20. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  21. Xu, S.; Qiu, X.; Wang, C.; Zhong, L.; Yuan, X. Desnet: Deep residual networks for Descalloping of ScanSar images. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 8929–8932. [Google Scholar]
  22. Wu, Z.; Shen, C.; Van Den Hengel, A. Wider or deeper: Revisiting the resnet model for visual recognition. Pattern Recognit. 2019, 90, 119–133. [Google Scholar] [CrossRef]
  23. Qiu, C.; Zhou, W. A survey of recent advances in CNN-based fine-grained visual categorization. In Proceedings of the 2020 IEEE 20th International Conference on Communication Technology (ICCT), Nanning, China, 28–31 October 2020; pp. 1377–1384. [Google Scholar]
  24. Ye, X.; Li, C.; Zhang, S.; Yang, P.; Li, X. Research on side-scan sonar image target classification method based on transfer learning. In Proceedings of the OCEANS 2018 MTS/IEEE Charleston, Charleston, SC, USA, 22–25 October 2018; pp. 1–6. [Google Scholar]
  25. Huo, G.; Wu, Z.; Li, J. Underwater object classification in sidescan sonar images using deep transfer learning and semisynthetic training data. IEEE Access 2020, 8, 47407–47418. [Google Scholar] [CrossRef]
  26. Yulin, T.; Jin, S.; Bian, G.; Zhang, Y. Shipwreck target recognition in side-scan sonar images by improved YOLOv3 model based on transfer learning. IEEE Access 2020, 8, 173450–173460. [Google Scholar] [CrossRef]
  27. Fuchs, L.R.; Gällström, A.; Folkesson, J. Object recognition in forward looking sonar images using transfer learning. In Proceedings of the 2018 IEEE/OES Autonomous Underwater Vehicle Workshop (AUV), Porto, Portugal, 6–9 November 2018; pp. 1–6. [Google Scholar]
  28. Zhang, H.; Tian, M.; Shao, G.; Cheng, J.; Liu, J. Target Detection of Forward-Looking Sonar Image Based on Improved YOLOv5. IEEE Access 2022, 10, 18023–18034. [Google Scholar] [CrossRef]
  29. Cheng, Z.; Huo, G.; Li, H. A Multi-Domain Collaborative Transfer Learning Method with Multi-Scale Repeated Attention Mechanism for Underwater Side-Scan Sonar Image Classification. Remote Sens. 2022, 14, 355. [Google Scholar] [CrossRef]
  30. Yang, X.; Zhang, H.; Qi, G.; Cai, J. Causal attention for vision-language tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9847–9857. [Google Scholar]
  31. Chen, Y.; Kalantidis, Y.; Li, J.; Yan, S.; Feng, J. A2-nets: Double attention networks. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS 2018), Montréal, QC, Canada, 3–8 December 2018. Volune 31. [Google Scholar]
  32. Lin, T.Y.; RoyChowdhury, A.; Maji, S. Bilinear CNN models for fine-grained visual recognition. In Proceedings of the IEEE International Conference on Computer Vision, Las Condes, Chile, 11–18 December 2015; pp. 1449–1457. [Google Scholar]
  33. Benarab, C.E.; Gui, S. CNN-Trans-Enc: A CNN-Enhanced Transformer-Encoder On Top Of Static BERT representations for Document Classification. arXiv 2022, arXiv:2209.06344. [Google Scholar]
  34. Luo, X.; Qin, X.; Wu, Z.; Yang, F.; Wang, M.; Shang, J. Sediment classification of small-size seabed acoustic images using convolutional neural networks. IEEE Access 2019, 7, 98331–98339. [Google Scholar] [CrossRef]
  35. Qin, X.; Luo, X.; Wu, Z.; Shang, J. Optimizing the sediment classification of small side-scan sonar images based on deep learning. IEEE Access 2021, 9, 29416–29428. [Google Scholar] [CrossRef]
  36. Gerg, I.D.; Monga, V. Structural prior driven regularized deep learning for sonar image classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–16. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed 3MATR model.
Figure 1. Framework of the proposed 3MATR model.
Electronics 12 00791 g001
Figure 2. Resnet local operation structure diagram.
Figure 2. Resnet local operation structure diagram.
Electronics 12 00791 g002
Figure 3. Detailed architecture diagram of the ResNet152 network.
Figure 3. Detailed architecture diagram of the ResNet152 network.
Electronics 12 00791 g003
Figure 4. CAMCL calculation flow chart.
Figure 4. CAMCL calculation flow chart.
Electronics 12 00791 g004
Figure 5. LAMCC calculation flow chart.
Figure 5. LAMCC calculation flow chart.
Electronics 12 00791 g005
Figure 6. Side-scan sonar dataset samples. (a) Three classes of side-scan image targets; (b) Sample distribution diagram.
Figure 6. Side-scan sonar dataset samples. (a) Three classes of side-scan image targets; (b) Sample distribution diagram.
Electronics 12 00791 g006
Figure 7. Image transformation result graph: (a) image sample; (b) center crop; (c) left bottom crop; (d) left top crop; (e) right bottom crop; (f) right top crop; (g) equal height stretch; (h) equal width stretch; (i) contrast transformation gamma = 0.87; (j) contrast transformation gamma = 1.07; (k) rotation = 45°; (l) rotation = 90°; (m) rotation = 135°; (n) rotation = 180°; (o) rotation = 225°; (p) rotation = 270°; (q) rotation = 315°; (r) left and right flip.
Figure 7. Image transformation result graph: (a) image sample; (b) center crop; (c) left bottom crop; (d) left top crop; (e) right bottom crop; (f) right top crop; (g) equal height stretch; (h) equal width stretch; (i) contrast transformation gamma = 0.87; (j) contrast transformation gamma = 1.07; (k) rotation = 45°; (l) rotation = 90°; (m) rotation = 135°; (n) rotation = 180°; (o) rotation = 225°; (p) rotation = 270°; (q) rotation = 315°; (r) left and right flip.
Electronics 12 00791 g007
Figure 8. Datasets used in the experiments: (a) Grayscale optical image samples; (b) SAR image samples; (c) SSS image samples.
Figure 8. Datasets used in the experiments: (a) Grayscale optical image samples; (b) SAR image samples; (c) SSS image samples.
Electronics 12 00791 g008
Figure 9. SSS target recognition results with DenseNet series networks: (a) DenseNet201 loss and accuracy curves; (b) DenseNet121 loss and accuracy curves; (c) DenseNet169 loss and accuracy curves.
Figure 9. SSS target recognition results with DenseNet series networks: (a) DenseNet201 loss and accuracy curves; (b) DenseNet121 loss and accuracy curves; (c) DenseNet169 loss and accuracy curves.
Electronics 12 00791 g009
Figure 10. SSS target recognition results of ResNet series networks: (a) ResNet50 loss and accuracy curves; (b) ResNet101 loss and accuracy curves; (c) ResNet152 loss and accuracy curves.
Figure 10. SSS target recognition results of ResNet series networks: (a) ResNet50 loss and accuracy curves; (b) ResNet101 loss and accuracy curves; (c) ResNet152 loss and accuracy curves.
Electronics 12 00791 g010
Figure 11. SSS target recognition results of VGGNet series networks: (a) VGGNet16 loss and accuracy curves; (b) VGGNet19 loss and accuracy curves.
Figure 11. SSS target recognition results of VGGNet series networks: (a) VGGNet16 loss and accuracy curves; (b) VGGNet19 loss and accuracy curves.
Electronics 12 00791 g011
Figure 12. SSS target transfer recognition results with ResNet series networks: (a) ResNet152 loss and accuracy curves; (b) ResNet101 loss and accuracy curves.
Figure 12. SSS target transfer recognition results with ResNet series networks: (a) ResNet152 loss and accuracy curves; (b) ResNet101 loss and accuracy curves.
Electronics 12 00791 g012
Figure 13. SSS target transfer recognition results with DenseNet series networks: (a) DenseNet121 loss and accuracy curves; (b) DenseNet169 loss and accuracy curves.
Figure 13. SSS target transfer recognition results with DenseNet series networks: (a) DenseNet121 loss and accuracy curves; (b) DenseNet169 loss and accuracy curves.
Electronics 12 00791 g013
Figure 14. SSS target migration recognition results with VGGNet16 network and the proposed method. (a) VGGNet16 loss and accuracy curves; (b) proposed method loss and accuracy curves.
Figure 14. SSS target migration recognition results with VGGNet16 network and the proposed method. (a) VGGNet16 loss and accuracy curves; (b) proposed method loss and accuracy curves.
Electronics 12 00791 g014
Figure 15. Forward-looking sonar dataset samples diagram.
Figure 15. Forward-looking sonar dataset samples diagram.
Electronics 12 00791 g015
Table 1. Resnet152 network parameter settings.
Table 1. Resnet152 network parameter settings.
Layer NameOutput SizeResnet152
conv1112 × 1127 × 7.64, s = 2
c o n v 2 _ x ( x = 1 3 ) 56 × 563 × 3 max pool, s = 2
1 × 1.64
3 × 3.64
1 × 1.256
c o n v 3 _ x ( x = 1 8 ) 28 × 281 × 1.128
3 × 3.128
1 × 1.512
c o n v 4 _ x ( x = 1 36 ) 14 × 141 × 1.256
3 × 3.256
1 × 1.1024
c o n v 5 _ x ( x = 1 3 ) 7 × 71 × 1.512
3 × 3.512
1 × 1.2048
Output1 × 1average pool
1000-dfc, softmax
Table 2. Specific data allocation for each category.
Table 2. Specific data allocation for each category.
CategoriesPlaneShipOther
Numbers
Total66484578
After dataset division (train set 70%, test set 30%)
Train46338404
Test20146174
Table 3. Comparison of different DL models for SSS image object recognition.
Table 3. Comparison of different DL models for SSS image object recognition.
MethodsOptimal OA (%)
DenseNet20195.02
DenseNet12195.13
DenseNet16996.15
ResNet5091.28
ResNet10196.41
ResNet15295.13
VGGNet1692.56
VGGNet1992.56
Table 4. Comparison of different methods for SSS image object recognition.
Table 4. Comparison of different methods for SSS image object recognition.
MethodsOptimal OA (%)
Shallow CNN [34]83.19
GoogleNet [35]91.86
VGG11 fine-tuning + semisynthetic data [24]92.51
VGG19 fine-tuning [25]94.67
VGG19 fine-tuning + semisynthetic data [25]97.76
SPDRDL [36]97.38
Our Methods98.72
Table 5. Comparison of different backbones for SSS image object recognition.
Table 5. Comparison of different backbones for SSS image object recognition.
Backbone NetworksOptimal OA (%)
ResNet1891.86
ResNet5093.5
ResNet15298.21
ResNet10198.46
DenseNet12196.41
DenseNet16991.54
DenseNet20189.23
VGGNet1698.46
VGGNet1994.67
AlexNet94.14
GoogleNet94.46
Proposed98.72
Table 6. Ablation experiments for backbone network trained on different data sets.
Table 6. Ablation experiments for backbone network trained on different data sets.
BackboneConv1Conv3_xConv5_xOptimal
Datasets Conv2_xConv4_xCAMCL LAMCCOA (%)
ImageNetImageNetSSS58.97
ImageNet/SAR/SSSSARSARSSS77.95
DatasetsImageNetSARSSS96.92
SARImageNetSSS98.72
Table 7. CAMCL and LAMCC ablation experiments results.
Table 7. CAMCL and LAMCC ablation experiments results.
Attention ModuleCAMCLLAMCCOptimal OA(%)
Backbone
97.69
ResNet152 97.18
98.72
Table 8. Comparison of different DL models for FLS image object recognition.
Table 8. Comparison of different DL models for FLS image object recognition.
MethodsOptimal OA (%)
DenseNet20189.07
DenseNet12188.87
DenseNet16989.91
ResNet5089.49
ResNet10188.14
ResNet15288.03
VGGNet1690.63
VGGNet1985.22
Proposed93.96
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, J.; Li, H.; Huo, G.; Li, C.; Wei, Y. Multi-Mode Channel Position Attention Fusion Side-Scan Sonar Transfer Recognition. Electronics 2023, 12, 791. https://doi.org/10.3390/electronics12040791

AMA Style

Wang J, Li H, Huo G, Li C, Wei Y. Multi-Mode Channel Position Attention Fusion Side-Scan Sonar Transfer Recognition. Electronics. 2023; 12(4):791. https://doi.org/10.3390/electronics12040791

Chicago/Turabian Style

Wang, Jian, Haisen Li, Guanying Huo, Chao Li, and Yuhang Wei. 2023. "Multi-Mode Channel Position Attention Fusion Side-Scan Sonar Transfer Recognition" Electronics 12, no. 4: 791. https://doi.org/10.3390/electronics12040791

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop