Next Article in Journal
A Review of Document Image Enhancement Based on Document Degradation Problem
Next Article in Special Issue
Adaptive Marginal Multi-Target Bayes Filter without Need for Clutter Density for Object Detection and Tracking
Previous Article in Journal
An Optimized Hybrid Transformer for Enhanced Ultra-Fine-Grained Thin Sections Categorization via Integrated Region-to-Region and Token-to-Token Approaches
Previous Article in Special Issue
Radar Target Localization with Multipath Exploitation in Dense Clutter Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Target Recognition via Cayley-Klein Measure and Shape Prior Information in Hyperspectral Imaging

1
Aviation Operations and Service Institute, Naval Aviation University, Yantai 264000, China
2
Cost Defence College, Naval Aviation University, Yantai 264000, China
3
School of Geosciences and Info-Physics, Central South University, Changsha 410083, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2023, 13(13), 7854; https://doi.org/10.3390/app13137854
Submission received: 4 April 2023 / Revised: 15 June 2023 / Accepted: 30 June 2023 / Published: 4 July 2023
(This article belongs to the Special Issue Object Detection Technology)

Abstract

:
Underwater target detection plays a vital role in various application scenarios, ranging from scientific research to military and industrial operations. In this paper, a detection method via the Cayley–Klein measure and a prior information of shape is proposed for the issue of hyperspectral underwater target identification. Firstly, by analyzing the data features of underwater targets and backgrounds, a background suppression algorithm based on Cayley–Klein measure is developed to enhance the differentiation between underwater targets and backgrounds. Then, a local peak-based algorithm is designed to discriminate potential underwater target points based on the local peak features of underwater targets. Finally, pseudo-target points are eliminated based on the priori shape information of underwater targets. Experiments show that the algorithm proposed is efficient and can effectively detect underwater targets from hyperspectral images.

1. Introduction

Generally, the process of classical method of target detection for hyperspectral remote sensing images can be expressed as follows: the spectral features of the image data points are projected onto a given plane by a certain criterion (i.e., detector) to ensure that the target and the backgrounds are located at different positions in this detection result plane, and then the target is separated from the background by a method of threshold segmentation [1]. During the detection process of hyperspectral targets, different models describing the background lead to different classifications of hyperspectral target detection methods. The classical methods of target detection for hyperspectral remote sensing images are divided into two categories: structured and unstructured backgrounds [2]. Three mathematical models exist to describe the spectral variation of pixels in hyperspectral images for structured background target detection: the probability density model, the subspace model, and the linear spectral hybrid model. Apart from the structured background model, a small number of hybrid detectors that combine the advantages of both have been investigated, and have been developed for both hybrid structured and unstructured detectors. Unlike the structured background method, the unstructured background detectors do not adopt this structured model of endmember signals with corresponding components to express background information. Instead, the background is treated as a statistical model conforming to a multivariate Gaussian distribution, and uniformly models the background and noise as the background. Typical methods include Generalized Likelihood Ratio (GLR), Adaptive Coherence Estimator (ACE), and Adaptive Matching Filter (AMF).
In recent years, several new methods of pattern recognition and machine learning have emerged in the field of hyperspectral remote sensing image target detection, including the Kernel Method [3], Sparse Representation Model [4], Discriminant Subspace Analysis [5,6] and Deep Learning Model [7]. For instance, a hyperspectral image anomaly detection method based on Support Vector Data Description (SVDD) was proposed by Li et al. [8], which can describe the image background depending on a few pixels without requiring the prior knowledge of data distribution. However, the shortcoming of this method mainly lies in the fact that the selection of kernel parameters of the algorithm is too dependent on human experience, and no scientific method of parameter selection is proposed. Wu et al. proposed a Collaborative Representation based Detector (CRD) algorithm for anomaly detection in hyperspectral images [9], assuming that background pixels can be expressed as linear combinations of neighboring pixels, while anomalous pixels do not follow this property, and using this property for anomalous target detection. The algorithm is a non-parametric method that does not require estimation of the background covariance matrix and has a closed solution.
Although algorithms such as deep learning have been progressively applied to hyperspectral target extraction, they also suffer from several bottlenecks: Firstly, the establishment of models in deep learning methods often requires certain numbers of reliable training samples, which is generally not satisfied in the field of hyperspectral target detection since the prior knowledge of hyperspectral target detection problem is only a single target spectrum, and the generalization and migration of the model is poor [10]. Secondly, it also possesses numerous parameters for deep learning methods to determine, and there is not enough data for cross-validation of multiple parameters in hyperspectral target detection [11]. Lastly, the computational complexity of machine learning methods tends to increase dramatically along with the expansion of problem scale, and the sample sets in hyperspectral target detection often reach megapixels, making it difficult to meet the requirements in real-time.
The above analysis shows that the existing hyperspectral target detection algorithms are mainly oriented to land domain scenarios, and there are few research cases for underwater target detection. Underwater target detection is a very interesting problem: many authors have proved that hyperspectral imagery could enable to retrieve efficiently bathymetry, coral reefs and sea grass [12,13,14,15]. However, only a few ones have dealt with the detection problem in such a context. For example, in [16], the purpose of the authors is more to classify pixels rather than to detect spared target pixels among the background. Most similarity criteria of hyperspectral pixels consider a single spectral dimension, and multi-source information including global, local, spectral, shape and other information is less comprehensive consideration. Afterward, it is difficult for the existing hyperspectral target detection algorithms to strike a balance between efficiency and accuracy, and further research is still required to improve the detection efficiency while meeting the accuracy requirements.
In this paper, we try to present a new method of detecting underwater targets in hyperspectral imaging, which not only guarantees the detection probability but also considers the efficiency. The development of the new algorithm is based on the hyperspectral images of underwater targets acquired by an unmanned airborne hyperspectral camera.

2. Study Area and Data

The hyperspectral remote sensing monitor employed in this experiment is the HY-9010-U unmanned aerial vehicle (UAV)-mounted hyperspectral imaging equipment, which was developed and produced by Hangzhou Spectral Imaging Technology Co., Ltd. (Hangzhou, China), as shown in Figure 1a, the details can be seen in Table 1. The site of data collection is near the coast of Yantai, with favorable aquatic conditions. The seawater in this area is 1–20 m in depth, crystal-clear, and free from various obstructions such as high-rise buildings. Also, the flat terrain along the coast is conducive to taking off and landing of the UAV. During the experiment, the weather was sunny, and the lighting conditions were good, ranging from 700 to 895 lux, which was conducive to the collection of hyperspectral images by the unmanned aerial vehicle. The data were collected on 9 October 2022, by deploying the UAV that mounted with a hyperspectral camera to perform aerial detection of the sea surface, the unmanned underwater vehicle (UUV) (Figure 1b) was used as the underwater loading platform, and the acquired hyperspectral data of underwater targets are shown in Figure 1c. The experimental area is an aquaculture zone where suspended particles in the water can reduce light penetration and may have an adverse effect on the experiment.
The UUV shown in Figure 1b is approximately 50 cm in length and 50 cm in width. As can be seen from Figure 1c, the channels of R, G, and B correspond to the 111th, 60th, and 32nd bands, respectively. The red area is the real reference of the UUV on the ground. To evaluate the impact of the UUV on target detection performance at different water depths, the hyperspectral dataset at four different water depths: 0.1 m, 1.5 m, 3.0 m, and 5 m were collected in this study. The number of samples at different depths is listed in Table 2. Light waves of different wavelengths are absorbed to varying degrees when penetrating water, showing a significant variation in the intensity of absorption. Since there is a valley in the absorption spectrum of water at 300 nm to 900 nm, it is favorable for light waves of this wave band to penetrate water, while those of other wave bands cannot penetrate due to the absorption by water.

3. Method of Recognition

3.1. Hyperspectral Background Field Suppression Based on Cayley–Klein Measure

Let I R d 1 × d 2 × d 3 be a hyperspectral image, where I refers to the original data, while d 1 , d 2 , and d 3 represent the height, width, and number of spectral bands of the hyperspectral image, respectively. Let x R d 3 be the spectral vector corresponding to each pixel. A number ( N ) of samples are randomly selected from the original hyperspectral image to comprise a dataset of x k I , where k = 1,2 , , N . Let m and represent the mean and covariance matrix of the dataset, respectively. In this study, given the extremely small proportion of the target of the UUV in the entire image, it can be approximated that x k represents the background field dataset. To calculate the discriminability of any spectral vector and the background field, the Mahalanobis distance is taken as a measure of similarity [17]. The Mahalanobis distance is expressed as follows:
d M x o , x p = x o x p T x o + x p
where d M represents the Mahalanobis distance measure, x o and x p are referred to as the two spectral vectors, and T denotes the transpose. The Mahalanobis distance plays a role in mitigating the impact of the background field, enhancing the discriminability between the target and the background field, and improving the accuracy of target detection. In addition, the Mahalanobis distance can be used to transform the high-dimensional space in an image into a low-dimensional space, thus improving the efficiency of data processing. The calculation of the Mahalanobis distance requires the inversion of the covariance matrix, which is in essence a linear measure. However, hyperspectral image data usually have high dimensionality, and the estimated covariance matrix is ill-conditioned in high-dimensional space, thus making its inversion unstable or impractical. Previous studies show that the Cayley–Klein metric matrix is capable to approximate the Mahalanobis distance in extreme cases [18]. Thus, this study introduces the Cayley–Klein measure to define the pixel similarity of the hyperspectral image. In order to obtain the Cayley–Klein measure, m and are used in this study to define two invertible symmetric matrices G ± as follows:
G ± = m m T m T m ± k 2 k > 0
where G + and G denote positive definite matrix and positive indefinite matrix, respectively. The bilinear form of G ± is expressed as follows:
σ ± x o , x p = x o T , 1 G ± x p 1 = x o m T x p m ± k 2 k > 0
According to Equation (3), Cayley–Klein measures can be combined such as d E ( x o , x p ) , d H ( x o , x p ) :
d E x o , x p = k 2 i log σ x o x p + + σ x o x p + 2 σ x o x o + σ x p x p + σ x o x p + σ x o x p + 2 σ x o x o + σ x p x p +
d H x o , x p = k 2 log σ x o x p + σ x o x p 2 σ x o x o σ x p x p σ x o x p σ x o x p 2 σ x o x o σ x p x p
According to experimental experience, d E ( x o , x p ) is conducive to the processing of hyperspectral image and the Cayley–Klein measure is selected with the value of k being set to 0.5. For the subsequent judgment of the underwater target, it is necessary to take the neighboring pixels around a certain pixel for similarity discrimination. At this time, m and can be obtained by using the local pixel set. To make use of both global and local information simultaneously, a new measure of d F x o , x p is proposed as follows:
d F x o , x p = 1 d E g l o b a l x o , x p d E l o c a l x o , x p
In Equation (6), d E g l o b a l x o , x p and d E l o c a l x o , x p represent the Cayley–Klein measures created by using the global and local sample sets, respectively. Global sample set are randomly selected in the whole hypespectral image, while local sample set are randomly selected within the local area of the target pixel. For more information, we recommend the reader to [19,20].
Figure 2 shows the comparative results of the three measures, and it can be found out that d E g l o b a l performs better than d E l o c a l in discriminating the target from the background. Compared with d E g l o b a l and d E l o c a l , d F shows greater homogeneity in the target area, while maintaining the discrimination between the target and background. This results in a good balance between inter-class and intra-class dispersion.
To address the impact of random sampling on the algorithm, the sampling process was repeated multiple times, with the average of the Cayley–Klein measure as the final result.

3.2. Local Peak Detection Algorithm for Target Points

Underwater robots and water bodies exhibit different features in hyperspectral imaging, which means local peaks appear in some particular regions. Allowing for this, a local peak detection algorithm is applied to screen out the pixels that may belong to the underwater autonomous vehicle. Let C R d 1 × d 2 × 1 be the Cayley–Klein measurement matrix corresponding to the hyperspectral image I R d 1 × d 2 × d 3 . For any point of x , y in the measurement matrix, the measurement of C x , y can be decomposed into:
C x , y = c t x , y + c b x , y + ε x , y
where c t , c b , and ε represent the target component, background component and noise, respectively. Assuming that there is a spectral difference between the UUV and the water, and there is supposed to be a local peak in the corresponding Cayley–Klein measurement, which is similar to the features of the point spread function (SPF), as can be described by a two-dimensional Gaussian function. For the local bright spot-shaped targets, the gray values in the locality are expected to show a peak that can be roughly described by a two-dimensional Gaussian distribution. That is, the target component of c t ( x , y ) can be approximately expressed as:
f t ( x , y ) = I i exp { 1 2 ( x x o i ) 2 1 i 2 +   ( y y o i ) 2 2 i 2 } , ( x , y ) Ω i 0 , ( x , y ) Ω i
In the above equation, Ω = Ω 1 Ω 2 Ω 3 Ω n , where Ω i represents the i -th target area, n indicates the number of target areas, and I i denotes the Cayley–Klein measurement of the i -th target. For those small targets, their small size can be expressed as any Δ x 2 + Δ y 2 = w 2 , satisfying x o i + Δ x , y o i + Δ y Ω i , where w represents a slightly larger value than the target size. By taken account into Equation (6), it can be known that:
F ( x a i , y a i ) F ( x a i + Δ x , y a i + Δ y ) = f t ( x o i , y o i ) + f b ( x o i , y o i ) f b ( x o i +   Δ x , y o i + Δ y ) + n ( x o i , y o i ) n ( x o i +   Δ x , y o i + Δ y ) , Δ x 2 + Δ y 2 = w 2
When the background changes at a relatively slow pace and the intensity of the noise is much lower than that of the target, the last two terms on the right-hand side of Equation (7) can be ignored, which means:
F ( x a i , y a i ) F ( x a i + Δ x y a i + Δ y ) = f t ( x o i , y o i ) = I i
Notably, the intensity level of noise must be lower compared with the target to satisfy the requirements of Equation (8). In case of salt-and-pepper noise, additional processing is required to apply the equation, which will be explained later.
The target must have a certain intensity, that is, I i > T , where T indicates a fixed threshold value. Therefore, the discriminant formula used for the selection of the peak features is expressed as:
F ( x o i , y o i ) F ( x o i + Δ x , y o i + Δ y ) > T
In practice, noise is inevitably present in the Cayley–Klein measures, which causes the frequent deviations of local peaks. To reduce the impact of noise on local peak detection, it is proposed in this paper to denoise the Cayley–Klein measures through median filtering. Figure 3 shows the local peak detection algorithm intended for the target points. It can be seen that there is a local peak of d F in the target area rather than the background area, and there are multiple d F peaks in the surrounding areas. On this basis, it is possible to distinguish the target from the background areas.

3.3. Pseudo-Target Criterion Based on Shape Prior Information

To distinguish the real underwater targets from the pseudo-targets as previously detected, the prior shape information of underwater targets is used to remove the pseudo-target points that contradict the prior shape information. The homogeneous region at any point is extracted from the local region of d F , which can be achieved through methods such as binarization and clustering.
The methods of thresholding and clustering can be used to extract homogenous regions. The thresholding method is simple to operate, but its threshold value is difficult to determine. An improper threshold value will affect the accurate extraction of homogeneous regions, resulting in difficulties in the subsequent shape discrimination. Clustering is a common unsupervised method, which divides a group of objects into several categories, with the core concept of maximizing the inter-class distances while minimizing the intra-class distances [21]. The classical clustering algorithms include: partitioning [22], density [23], graph theory [19], and grid-based method [24]. The K-means [25] is a typical partitioned clustering algorithm, which is only applicable to spherical clustering, and the clustering results are easily affected by the initial clustering center. The density-based noise can be found by using the Density-Based Spatial Clustering of Application with Noise (DBSCAN) [26] algorithm, but the clustering results are easily affected by the threshold and neighborhood radius. Rodriguez et al. [27] proposed the Density Peak Clustering (DPC) algorithm, which believes that the density of the cluster center is higher than that of the surrounding data points and is far away from other high-density points. However, the DPC algorithm only calculates the density according to the distance between data points, which is not suitable for datasets with complex shapes and densities. When distributing data points based on distances, the robustness of the algorithm is poor, and the clustering results are sensitive to the value of the cut-off distance.
The Local Gravitation Model (LGM) [28] is a method for data clustering. It is different from the traditional gravity clustering model, where the modulus length and direction information of the resultant force of the local gravitation are used in the calculation of the local center measure. Considering the efficiency and accuracy of the algorithm, the local gravity gravitation clustering model [29] is used to extract the homogeneous regions from the measurements. The principle of the LGM clustering algorithm is as follows:
Inspired by Newton’s law of universal gravitation, the local gravitation of data points in the clustering process can reflect the relationship between data points and their neighbors. According to the law of universal gravitation, the attraction between two particles can be defined as:
F i j = G m i m j d i j 2 D i j
where F i j represents the gravitation between the points of X i and X j , G represents the coefficient of gravitation, m represents the mass of the point X i , d i j represents the distance between the points of X i and X j , the unit vector D i j is the direction of the line between the two points. Since the gravitation in the local neighborhood is calculated, there is no significant difference in the distance between the data points and the neighbors, so Equation (12) can be simplified as
F i j = G m i m j D i j
The Local Resultant Force (LRF) on X i produced by its k nearest neighbor points can be simplified as:
L R F i , k = j = 1 k F i j = G m i j = 1 k m j D i j
where k represents the number of the nearest neighbor points of X i . According to Equation (14), the greater the mass of X i , the greater the influence on the neighborhood. Similarly, the smaller the mass, the more susceptible it is to the influence of the nearest neighbor points. Therefore, the LRF can be simplified as:
L R F i , k = 1 m i j = 1 k D i j s . t .   m i = 1 / j = 1 k d i j
On the basis of the above equations, the LGM clustering process includes the following: Firstly, the concentration of the data points is calculated using the LGM model, and the distance between each data point and the points with high concentration is determined based on the concentration. Secondly, the data points with high concentration and high distance values are selected as the clustering centers. Finally, the rest of the data points are allocated based on the principle that the concentration of the points inside the cluster is much higher than that of the boundary points, and the parameters are automatically adjusted by balancing the K-Nearest Neighbor.
The area of the homogeneous region at a point is denoted as A i , which is supposed to satisfy the following conditions:
T A 1 < A i < T A 2
where T A 1 and T A 2 represent the lower and upper limits of the area threshold, respectively. Given a homogeneous region, the length of w 1 and width of w 2 of its minimum bounding rectangle can be obtained. Due to the impact of various factors, there may be some true target areas missing from the homogeneous regions, but the proportion of true extractions is supposed to exceed a certain limit of T h .
A i w 1 × w 2 > T h
Since the underwater targets, which are treated as the focus of this study are squares, the aspect ratio of the minimum bounding rectangle should be close to 1. However, due to the inability to extract the true target area completely, the threshold T l of the aspect ratio can be appropriately relaxed.
w 1 w 2 < T l
Based on the prior experience, the value of T A 1 and T A 2 are set to 30 and 200, respectively; T h and T l are set to 0.4 and 1.6, respectively. When the conditions as set out in Equations (16)–(18) are not satisfied in the homogeneous region at a point, this point is treated as a pseudo-target point. The threshold values depend on the spatial resolution of the hyperspectral image. To make the threshold more applicable, we can apply the spatial resolution of the image used in this study as a reference, and the ratio of the resolution of new images to the reference image resolution can be used as a scaling factor for the threshold.
Figure 4 lists the shape information of the candidate target points in the three situations: Situation 1 indicates that although the area of the homogeneous region of the candidate target point meets the requirements of Equation (16), its aspect ratio is higher than T l , indicating that the candidate point is more inclined to linear features (such as cables), which is not consistent with the shape of the underwater target in our work. Situation 2 indicates that the homogeneous area of the candidate target point is smaller than the lower limit T A 1 of the area threshold, indicating that the candidate point tends to be a noise point, which is inconsistent with the area information of the underwater target in this work, and is not a potential underwater target point. Situation 3 indicates that the homogenous region of the candidate target points satisfies Equations (16)–(18), indicating that the candidate points are in good agreement with the shape of the underwater targets in this work, and can be judged as potential underwater target points.

4. Results and Analysis

Considering the particularity of the scene in this work and combining it with the existing literature, existing evaluation indexes are adopted and improved in this work to objectively evaluate the recognition accuracy of underwater targets. TP (True Positive), FP (False Positive), TN (True Negative), and FN (False Negative) are assumed as four indexes calculated from the confusion matrix, which can be used to calculate the False Negatives Rate (FNR) and the False Positive Rate (FPR).
FNR represents the ability of the target detection model to correctly predict the purity of negative samples and reduces the prediction of positive samples as negative samples, that is, the proportion of positive samples predicted as negative samples in the total positive samples. The smaller the FNR value, the better the target detection performance.
F N R = F N P
FPR represents the ability of the target detection model to correctly predict the purity of positive samples, and reduces the prediction of negative samples to positive samples, that is, the proportion of negative samples predicted as positive samples to the total number of negative samples. The smaller the FPR value, the better the target detection performance.
F P R = F P N
The Signal-to-Clutter Ratio (SCR) is defined in Equation (21), where the higher the SCR for the target, the easier it will be detected.
S C R = μ t μ b σ b
where μ t is the average pixel value of the target, μ b and σ b are the average pixel value and the standard deviation value of the pixel value for the adjacent area of the target, respectively.
Since the area of background is much larger than that of target in the research image, the number of the background and the target samples will be unbalanced, (i.e., the positive and negative samples are unbalanced), when all the background areas are used in accuracy evaluation, which reduces the efficiency of evaluating SCR and ROC (Receiver Operating Characteristic). In order to reduce the impact, the target window and background window are set to 20 × 20 pixels and 4 0 × 40 pixels, respectively, as shown in Figure 5. The resetting of the calculation area is just for the comparison of the performance of the algorithms.
In general, the FNR and FPR are global indicators, while the improved SCR and ROC in our work belong to local indicators. Therefore, the evaluation system in this work is reasonable by comprehensively evaluating the target detection accuracy from both the global and local areas.
In order to verify the performance and advantages of the proposed algorithms of underwater target detection, seven target detection algorithms for hyperspectral images are compared and analyzed, including Kernel Isolation Forest (KIF) [30], Visual Attention and Background Subtraction (VABS) [31], Sub-Features Grouping and Binary Accumulation (SFBA) [32], Spectral–Spatial Fusion Anomaly Detection Method (SSFAD) [33], Global Reed–Xiaoli (GRX) [34], Local Reed–Xiaoli (LRX) [35] and Low Rank and Sparse Representation (LRASR) [36]. Table 2 shows the comparison results of the precision in different algorithms.
Figure 6, Figure 7, Figure 8 and Figure 9 show the results of experiments 1 to 4, respectively. As can be seen from Figure 9b–i, the background regions of KIF, SFBA and SSFAD are much rougher than those of the algorithms in our work, including VABS, GRX, LRX and LRASA. Figure 6b–i show that the anomalous points detected by some algorithms are located at the central region of the target, while in some other algorithms the points are located at the edge region. For experiment 1, multiple peaks appear in both the target and background regions of SFBA, with the peak values of the target region lower than those in the background region, which leads to the misdetection in experiment 1.
At the same time, with the increase of the depth of the UUV, the background areas of KIF, SFBA and SSFAD become rougher, indicating that the increase in the depth of water will aggravate the mixed pixel phenomenon, which is unfavorable in the target detection. Meanwhile, Figure 9c’,g’ show that the background and target of VABS and LRX become indistinguishable at the water depth of 5 m, indicating a failure in detection efficiency. In addition, the size of the peak value of the target region detected by most algorithms begins to weaken with the increase of water depth, indicating that the distinction between the underwater target and water decreases, which is also detrimental to the recognition of underwater targets. The weakening of water depth on target detection performance is also verified in Table 3. For example, when the water depth is 0.1 m, all the algorithms successfully detect underwater targets except SFBA. However, when the water depth is 5 m, the VABS, SFBA, SSFAD, LRX, and LRASR all fail in the underwater target detection. Table 3 also shows that the algorithms proposed in our work achieve the best performance in the comparison of the FNR and FPR, and the SCR is also close to the best performance of similar algorithms, which demonstrates the advantages and strong detection performance of the proposed algorithm.
The test environment created in this work for a comparative analysis of the efficiency achieved by these algorithms is the Intel(R) Xeon(R) processor with a main frequency of 3.60 GHZ and an internal storage of 32 G. The software environment is MATLAB R2019b. To begin with, the computation is repeated five times for each experimental image, with the average time taken for each computation as the final result. As shown in Table 4, the algorithm proposed in this work leads to a significant improvement compared with those of Kernel IF, SSFAD, SRX_LOCAL and LRASR, which makes the processing faster. roughly over 16–60 times It is comparable with VABS, but slightly outperformed by SFBA and Global RX. This algorithm can be refined to further improve computational efficiency. As can be seen above, this algorithm improves accuracy on the basis of ensuring efficiency. Compared with other mainstream ones, the algorithm proposed in this work performs better in balancing accuracy and efficiency.

5. Conclusions

In this work, a detection method based on Cayley–Klein measure and a prior shape information is proposed for better reorganization of underwater targets, which leads to satisfactory outcomes in processing the hyperspectral images captured by UAV. Then, both experiments and comparative analysis are conducted to draw the following conclusions.
Firstly, the Mahalanobis distance can mitigate the impact of background on target identification. Meanwhile, the hyperspectral images are mapped from the high-dimensional space to the low-dimensional space, thus improving the efficiency of data processing. However, this distance is effectively a linear mapping, which means it can fail in high-dimensional space. To address the constraint on processing high-dimensional data with the Mahalanobis distance, we propose to use the Cayley–Klein measure to improve efficiency.
Secondly, the Cayley–Klein measure proposed in this study takes full advantage of global and local information to suppress the background, which enhances the discrimination between the target and the background while improving the heterogeneity of the target area. This is conducive to the identification of targets. Compared with traditional algorithms, the algorithm proposed in this work is applicable to detect the targets at a water depth of up to 5 m.
Thirdly, based on the prior shape information of the underwater target, a false target detection algorithm is developed to improve the accuracy of target identification, which reduces the impact of the false positive rate.
Lastly, the proposed algorithm achieves a high computational efficiency that is comparable with other mainstream target detection algorithms, and a good balance is reached between efficiency and accuracy.
There are limitations of the algorithms proposed in this work because only spectral information is used to conduct underwater target detection and given the account of properties of electromagnetic wave transmission in water. Therefore, it is worthwhile to further explore how to make use of the combined information about spectrum, the intensity of light, and polarization. Another research direction would be to research the relationship between the detected water depth of the underwater target and the water quality.

Author Contributions

Conceptualization, B.Z. and Z.M.; methodology, B.Z., F.Z. and X.L.; investigation, Y.S. and L.L.; writing—original draft preparation, P.L.; writing—review and editing, B.Z.; supervision, Z.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China grant number 62005318.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the editors and anonymous reviewers for their valuable comments that can greatly improve our manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, L. Advance and Future Challenges in Hyperspectral Target Detection. Geomat. Inf. Sci. Wuhan Univ. 2014, 39, 1387–1400. [Google Scholar]
  2. Nasrabadi, N.M. Hyperspectral target detection: An overview of current and future challenges. IEEE Signal Process. Mag. 2013, 31, 34–44. [Google Scholar] [CrossRef]
  3. Kwon, H.; Nasrabadi, N.M. Kernel matched subspace detectors for hyperspectral target detection. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 28, 178–194. [Google Scholar] [CrossRef]
  4. Zhu, D.; Du, B.; Zhang, L. Target dictionary construction-based sparse representation hyperspectral target detection methods. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1254–1264. [Google Scholar] [CrossRef]
  5. Sun, S.; Liu, J.; Sun, S. Hyperspectral Subpixel Target Detection Based on Interaction Subspace Model. Pattern Recognit. 2023, 139, 109464. [Google Scholar] [CrossRef]
  6. Zhou, Y.; Chen, P.; Liu, N.; Yin, Q.; Zhang, F. Graph-Embedding Balanced Transfer Subspace Learning for Hyperspectral Cross-Scene Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 2944–2955. [Google Scholar] [CrossRef]
  7. Hu, X.; Xie, C.; Fan, Z.; Duan, Q.; Zhang, D.; Jiang, L.; Wei, X.; Hong, D.; Li, G.; Zeng, X. Hyperspectral anomaly detection using deep learning: A review. Remote Sens. 2022, 14, 1973. [Google Scholar] [CrossRef]
  8. Li, K.; Ling, Q.; Qin, Y.; Wang, Y.; Cai, Y.; Lin, Z.; An, W. Spectral-spatial deep support vector data description for hyperspectral anomaly detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522316. [Google Scholar] [CrossRef]
  9. Wu, Z.; Su, H.; Tao, X.; Han, L.; Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Hyperspectral anomaly detection with relaxed collaborative representation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5533417. [Google Scholar] [CrossRef]
  10. Zhang, H.-K.; Li, Y.; Jiang, Y.-N. Deep Learning for Hyperspectral Imagery Classification: The State of the Art and Prospects. Acta Autom. Sin. 2018, 44, 961–977. [Google Scholar]
  11. Jia, S.; Jiang, S.; Lin, Z.; Li, N.; Xu, M.; Yu, S. A survey: Deep learning for hyperspectral image classification with few labeled samples. Neurocomputing 2021, 448, 179–204. [Google Scholar] [CrossRef]
  12. Lee, Z.; Carder, K.; Mobley, C.; Steward, R.; Patch, J. Hyperspectral remote sensing for shallow waters. ii. deriving bottom depths and water properties by optimization. Appl. Opt. 1999, 38, 3831–3843. [Google Scholar] [CrossRef] [Green Version]
  13. Adler-Golden, S.; Acharya, P.; Berk, A.; Matthew, M.; Gorodetzky, D. Remote bathymetry of the littoral zone from A VIRIS, LASH and QuickBird imagery. IEEE Trans. Geosci. Remote Sens. 2005, 43, 337–347. [Google Scholar] [CrossRef]
  14. Kallio, K.; Kutser, T.; Hannonen, T.; Koponen, S.; Pul-liainen, J.; Vepsäläinen, J.; Pyhälahti, T. Retrieval of water quality from airborne imaging spectrometry of various lake types in different seasons. Sci. Total Environ. 2001, 268, 59–77. [Google Scholar] [CrossRef] [PubMed]
  15. Vahtmäe, E.; Kutser, T.; Martin, G.; Kotta, J. Feasibility of hyperspectral remote sensing for mapping benthic macroalgal cover in turbid coastal waters—A baltic sea case study. Remote Sens. Environ. 2008, 101, 342–351. [Google Scholar] [CrossRef]
  16. David, G. An Underwater Target Detection Framework for Hyperspectral Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1798–1810. [Google Scholar]
  17. De Maesschalck, R.; Jouan-Rimbaud, D.; Massart, D.L. The Mahalanobis distance. Chemomeasures Intell. Lab. Syst. 2000, 50, 1–18. [Google Scholar] [CrossRef]
  18. Struve, H.; Struve, R. Projective spaces with Cayley-Klein metrics. J. Geom. 2004, 81, 155–167. [Google Scholar] [CrossRef]
  19. Bi, Y.; Fan, B.; Wu, F. Beyond Mahalanobis metric: Cayley-Klein metric learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2339–2347. [Google Scholar]
  20. Bi, Y.; Fan, B.; Wu, F. Multiple Cayley-Klein metric learning. PLoS ONE 2017, 12, e0184865. [Google Scholar] [CrossRef] [Green Version]
  21. Mika, S.; Ratsch, G.; Weston, J.; Scholkopf, B.; Mullers, K.-R. Fisher discriminant analysis with kernels. Neural Networks for Signal Processing IX. In Proceedings of the 1999 IEEE Signal Processing Society Workshop (Cat. No. 98th8468), Madison, WI, USA, 25 August 1999; pp. 41–48. [Google Scholar]
  22. Patibandla, R.S.M.; Veeranjaneyulu, N. A SimRank based ensemble method for resolving challenges of partition clustering methods. J. Sci. Ind. Res. 2022, 79, 323–327. [Google Scholar]
  23. Kriegel, H.; Kröger, P.; Sander, J.; Zimek, A. Density-based clustering. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2011, 1, 231–240. [Google Scholar] [CrossRef]
  24. Liu, Y.; Tu, W.; Zhou, S.; Liu, X.; Song, L.; Yang, X.; Zhu, E. Deep graph clustering via dual correlation reduction. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 7603–7611. [Google Scholar]
  25. Cheng, M.; Ma, T.; Ma, L.; Yuan, J.; Yan, Q. Adaptive grid-based forest-like clustering algorithm. Neurocomputing 2022, 481, 168–181. [Google Scholar] [CrossRef]
  26. Zhao, X.; Nie, F.; Wang, R.; Li, X. Improving projected fuzzy K-means clustering via robust learning. Neurocomputing 2022, 491, 34–43. [Google Scholar] [CrossRef]
  27. Ouyang, T.; Shen, X. Online structural clustering based on DBSCAN extension with granular descriptors. Inf. Sci. 2022, 607, 688–704. [Google Scholar] [CrossRef]
  28. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [Green Version]
  29. Wang, Z.; Yu, Z.; Chen CL, P.; You, J.; Gu, T.; Wong, H.-S.; Zhang, J. Clustering by local gravitation. IEEE Trans. Cybern. 2017, 48, 1383–1396. [Google Scholar] [CrossRef]
  30. Campbell, J.B.; Wynne, R.H. Introduction to Remote Sensing; Guilford Press: New York, NY, USA, 2011. [Google Scholar]
  31. Li, S.; Zhang, K.; Duan, P.; Kang, X. Hyperspectral anomaly detection with kernel isolation forest. IEEE Trans. Geosci. Remote Sens. 2019, 58, 319–329. [Google Scholar] [CrossRef]
  32. Xiang, P.; Song, J.; Qin, H.; Tan, W.; Li, H.; Zhou, H. Visual attention and background subtraction with adaptive weight for hyperspectral anomaly detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2270–2283. [Google Scholar] [CrossRef]
  33. Yuan, S.; Shi, L.; Yao, B.; Li, F.; Du, Y. A hyperspectral anomaly detection algorithm using sub-features grouping and binary accumulation. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6007505. [Google Scholar] [CrossRef]
  34. Hou, Z.; Cheng, S.; Hu, T. A spectral-spatial fusion anomaly detection method for hyperspectral imagery. arXiv 2022, arXiv:2202.11889. [Google Scholar]
  35. Guo, Q.; Zhang, B.; Ran, Q.; Gao, L.; Li, J.; Plaza, A. Weighted-RXD and linear filter-based RXD: Improving background statistics estimation for anomaly detection in hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2351–2366. [Google Scholar] [CrossRef]
  36. Xu, Y.; Wu, Z.; Li, J.; Plaza, A.; Wei, Z. Anomaly detection in hyperspectral images based on low-rank and sparse representation. IEEE Trans. Geosci. Remote Sens. 2015, 54, 1990–2000. [Google Scholar] [CrossRef]
Figure 1. (a) HY-9010-U UAV -mounted hyperspectral imaging instrument; (b) UUV; (c) The visible light image corresponding to the hyperspectral image where the red area indicates the location of UUV.
Figure 1. (a) HY-9010-U UAV -mounted hyperspectral imaging instrument; (b) UUV; (c) The visible light image corresponding to the hyperspectral image where the red area indicates the location of UUV.
Applsci 13 07854 g001
Figure 2. (a) The pseudo-color image corresponding to the hyperspectral image, where the red box indicates the underwater target region; (bd) show the results of d E g l o b a l , d E l o c a l , and d F , respectively; (eg) show the 3D views of d E g l o b a l , d E l o c a l , and d F , respectively.
Figure 2. (a) The pseudo-color image corresponding to the hyperspectral image, where the red box indicates the underwater target region; (bd) show the results of d E g l o b a l , d E l o c a l , and d F , respectively; (eg) show the 3D views of d E g l o b a l , d E l o c a l , and d F , respectively.
Applsci 13 07854 g002aApplsci 13 07854 g002b
Figure 3. The first column including (a,d,g) shows the false-color images corresponding to the hyperspectral images, (a) represents the target area where the red box indicates the underwater target region, and (d) represents the background area; The second column including (b,e,h) shows the d F results, and the third column including (c,f,i) shows the 3D view of d F .
Figure 3. The first column including (a,d,g) shows the false-color images corresponding to the hyperspectral images, (a) represents the target area where the red box indicates the underwater target region, and (d) represents the background area; The second column including (b,e,h) shows the d F results, and the third column including (c,f,i) shows the 3D view of d F .
Applsci 13 07854 g003
Figure 4. (a) The result of d F ; (b) The result of local gravity clustering corresponding to d F ; (c) The homogeneous region corresponding to the center pixel.
Figure 4. (a) The result of d F ; (b) The result of local gravity clustering corresponding to d F ; (c) The homogeneous region corresponding to the center pixel.
Applsci 13 07854 g004aApplsci 13 07854 g004b
Figure 5. Diagram of the target and background window size setting involved in SCR and ROC calculation.
Figure 5. Diagram of the target and background window size setting involved in SCR and ROC calculation.
Applsci 13 07854 g005
Figure 6. Results of experiment 1: (a) A red box indicating the location of the underwater target for test on the image of visible light in Region 1. (bi) show results of the algorithm proposed in this study, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively. (b’i’) show a partial 3D view of the surrounding region of the underwater target detected by the presented method, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively.
Figure 6. Results of experiment 1: (a) A red box indicating the location of the underwater target for test on the image of visible light in Region 1. (bi) show results of the algorithm proposed in this study, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively. (b’i’) show a partial 3D view of the surrounding region of the underwater target detected by the presented method, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively.
Applsci 13 07854 g006
Figure 7. Results of experiment 2: (a) A red box indicating the location of the underwater target for test on the image of visible light in Region 1. (bi) show results of the algorithm proposed in this study, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively. (b’i’) show a partial 3D view of the surrounding region of the underwater target detected by the presented method, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively.
Figure 7. Results of experiment 2: (a) A red box indicating the location of the underwater target for test on the image of visible light in Region 1. (bi) show results of the algorithm proposed in this study, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively. (b’i’) show a partial 3D view of the surrounding region of the underwater target detected by the presented method, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively.
Applsci 13 07854 g007
Figure 8. Results of experiment 3: (a) A red box indicating the location of the underwater target for test on the image of visible light in Region 1. (bi) show results of the algorithm proposed in this study, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively. (b’i’) show a partial 3D view of the surrounding region of the underwater target detected by the presented method, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively.
Figure 8. Results of experiment 3: (a) A red box indicating the location of the underwater target for test on the image of visible light in Region 1. (bi) show results of the algorithm proposed in this study, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively. (b’i’) show a partial 3D view of the surrounding region of the underwater target detected by the presented method, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively.
Applsci 13 07854 g008
Figure 9. Results of experiment 4: (a) A red box indicating the location of the underwater target for test on the image of visible light in Region 1. (bi) show results of the algorithm proposed in this study, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively. (b’i’) show a partial 3D view of the surrounding region of the underwater target detected by the presented method, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively.
Figure 9. Results of experiment 4: (a) A red box indicating the location of the underwater target for test on the image of visible light in Region 1. (bi) show results of the algorithm proposed in this study, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively. (b’i’) show a partial 3D view of the surrounding region of the underwater target detected by the presented method, Kernel IF, VABS, SFBA, SSFAD, Global RX, SRX_LOCAL and LRASR, respectively.
Applsci 13 07854 g009
Table 1. Parameters of the HY-9010-U UAV -mounted hyperspectral imaging instrument.
Table 1. Parameters of the HY-9010-U UAV -mounted hyperspectral imaging instrument.
Imaging ModeDispersive Push-Broom Mode
Spectral range400–1000 nm
Spectral resolution2.8 nm
Spatial resolution7 cm
Number of spatial pixels1920 (1×), 480 (4×)
Number of spectral bands1200 (1×), 300 (4×)
Slit25 μm
Relative apertureF2.6
Output pixel bit depth12 bits
Maximum frame rate (full spectrum acquisition)128 fps
Lens focal length12.5 mm/35 mm interchangeable.
Height of UAV100 m
Table 2. The number of the underwater target at different depths.
Table 2. The number of the underwater target at different depths.
Depth of the TargetNumber of the Samples
0.1 m65
1.5 m23
3 m17
5 m53
Table 3. Comparison results of the precision in different algorithms.
Table 3. Comparison results of the precision in different algorithms.
ImageIndicatorsProposed AlgorithmKernel IFVABSSFBASSFADGlobal RXSRX_LOCALLRASR
#1
Depth: 0.1 m
FNR00010000
FPR00.950.950.990.930.940.940.92
SCR0.395.23\0.824.06228.7126.82112.52
#2
Depth: 1.5 m
FNR00000010
FPR00.960.950.940.940.920.830.99
SCR4.954.3412.121.130.802.71\9.95
#3
Depth: 3.5 m
FNR00101111
FPR00.970.950.950.880.950.950.94
SCR3.660.10\0.120.080.140.47\
#4
Depth: 5.0 m
FNR00111011
FPR0.50.971110.9811
SCR1.590.80\0.040.04\\\
The symbol ‘\’ represents that the standard deviation value of the pixel value of the adjacent area of the target is 0 or does not exist, which leads to the failure of the calculation of SCR.
Table 4. Comparison of processing efficiency between different algorithms.
Table 4. Comparison of processing efficiency between different algorithms.
Experimental AreaRepeated TimesComputational Cost (s)
Proposed AlgorithmKernel IFVABSSFBASSFADGlobal RXSRX_LOCALLRASR
#1
Water depth: 0.1 m
1st55.11724.4088.853.261404.551.882421.822799.84
2nd53.84726.0996.753.071504.591.702439.653099.21
3rd32.06778.0690.033.401390.041.812355.662762.05
4th50.43807.5790.593.201454.751.742121.742814.04
5th40.73770.7897.463.231552.872.142488.193010.86
Mean value46.43761.3892.743.231461.361.852365.412897.20
Variance9.8135.754.050.1268.190.17144.26148.65
#2
Water depth: 1.5 m
1st40.50615.5177.933.011108.571.662002.572520.30
2nd36.58650.1480.002.991248.181.981970.782390.84
3rd43.65600.2765.133.091149.641.962112.072371.72
4th34.13705.4565.932.991335.022.201815.172384.53
5th36.49617.4276.433.071168.501.982045.392199.38
Mean value38.27637.7673.083.031201.981.961989.202373.35
Variance3.7841.987.020.0590.020.19110.75114.34
#3
Water depth: 3.0 m
1st82.76586.6967.533.011061.491.521602.402122.90
2nd84.46574.1161.413.021098.431.431607.271910.10
3rd85.83543.9780.643.30970.731.671656.372055.71
4th67.14547.1371.563.151020.241.851559.981952.88
5th78.78588.9376.663.011034.111.481599.742133.14
Mean value79.79568.1771.563.101037.001.591605.152034.95
Variance7.5521.437.550.1347.580.1734.30100.16
#4
Water depth: 5.0 m
1st91.83676.2992.233.171326.471.871926.322417.60
2nd82.01677.01101.583.101294.711.871834.172337.64
3rd89.96656.6387.333.501342.932.061711.872447.91
4th82.88673.3290.103.421155.071.521785.972020.00
5th92.32668.7798.833.201207.061.691855.562504.81
Mean value87.80670.4094.013.281265.251.801822.782345.59
Variance4.988.355.990.1780.900.2179.97191.75
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, B.; Zhang, F.; Sun, Y.; Li, X.; Liu, P.; Liu, L.; Miao, Z. Underwater Target Recognition via Cayley-Klein Measure and Shape Prior Information in Hyperspectral Imaging. Appl. Sci. 2023, 13, 7854. https://doi.org/10.3390/app13137854

AMA Style

Zhang B, Zhang F, Sun Y, Li X, Liu P, Liu L, Miao Z. Underwater Target Recognition via Cayley-Klein Measure and Shape Prior Information in Hyperspectral Imaging. Applied Sciences. 2023; 13(13):7854. https://doi.org/10.3390/app13137854

Chicago/Turabian Style

Zhang, Bin, Fan Zhang, Yansen Sun, Xiaojie Li, Pei Liu, Liang Liu, and Zelang Miao. 2023. "Underwater Target Recognition via Cayley-Klein Measure and Shape Prior Information in Hyperspectral Imaging" Applied Sciences 13, no. 13: 7854. https://doi.org/10.3390/app13137854

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop