Next Article in Journal
Early Yield Forecasting of Maize by Combining Remote Sensing Images and Field Data with Logistic Models
Next Article in Special Issue
AI-TFNet: Active Inference Transfer Convolutional Fusion Network for Hyperspectral Image Classification
Previous Article in Journal
Block-Scrambling-Based Encryption with Deep-Learning-Driven Remote Sensing Image Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Transfer-Based Framework for Underwater Target Detection from Hyperspectral Imagery

1
School of Electronics and Information, Northwestern Polytechnical University, Xi’an 710100, China
2
Military Representative Office of Naval Equipment Department for Luoyang Region, Luoyang 471003, China
3
Luoyang Electronic Equipment Test Center—LEETC, Luoyang 471003, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(4), 1023; https://doi.org/10.3390/rs15041023
Submission received: 26 December 2022 / Revised: 31 January 2023 / Accepted: 9 February 2023 / Published: 13 February 2023
(This article belongs to the Special Issue Active Learning Methods for Remote Sensing Data Processing)

Abstract

:
The detection of underwater targets through hyperspectral imagery is a relatively novel topic as the assumption of target background independence is no longer valid, making it difficult to directly detect underwater targets using land target information. Meanwhile, deep-learning-based methods have faced challenges regarding the availability of training datasets, especially in underwater conditions. To solve these problems, a transfer-based framework is proposed in this paper, which exploits synthetic data to train deep-learning models and transfers them to real-world applications. However, the transfer becomes challenging due to the imparity in the distribution between real and synthetic data. To address this dilemma, the proposed framework, named the transfer-based underwater target detection framework (TUTDF), first divides the domains using the depth information, then trains models for different domains and develops an adaptive module to determine which model to use. Meanwhile, a spatial–spectral process is applied prior to detection, which is devoted to eliminating the adverse influence of background noise. Since there is no publicly available hyperspectral underwater target dataset, most of the existing methods only run on simulated data; therefore, we conducted expensive experiments to obtain datasets with accurate depths and use them for validation. Extensive experiments verify the effectiveness and efficiency of TUTDF in comparison with traditional methods.

1. Introduction

Hyperspectral imagery (HSI) contains data in the form of a cube containing two spatial dimensions and one spectral dimension. Unlike RGB images, which only contain three bands, each pixel in HSI has rich spectral information, which usually possesses hundreds of immediately adjacent bands. Generally, the reflection spectrum of a material is called its spectral signature, which could be used to distinguish this from others. Thus, HSI is widely exploited in various fields, such as agricultural management, resource exploration, and military detection. Research on HSI usually includes dimensionality reduction [1], target detection [2,3] and classification [4], among which hyperspectral target detection (HTD) is a significant part. This can be divided into object-level and pixel-level detection. Object-level detection [5] is the detection of a specific target object as a whole, and existing methods tend to detect this target by extracting spatial–spectral features and using RGB image object-detection algorithms. However, pixel-level detection is more mainstream. This is usually regarded as a binary classifier that can be used to determine whether the pixel belongs to the target or background [6]. To be more precise, differing from the binary classifier, it performs a classification by assigning each pixel to one of two classes: the detector used to perform HTD finds an appropriate threshold, which corresponds to a boundary between the class of target absence and the class of target presence [4]. Generally, detectors fall into three categories according to whether their background information and target spectral information are known. In the first category, neither is available—for example, RX [7], which merely detects anomalies with limited performance. The target information is known in the second group, while the background information is unknown, as in CEM [8]. As for the third category, such as OSP [9], both the target and background information are known.
In general, the spectral signature can directly reflect the type of substance; there seems to be no need to learn and classify them using a network. However, when the target is submerged and the situation is not the same, the signature of the to-be-detected target will change due to interference from the water background. Compared with land-based HTD, less research has been conducted on underwater target detection (UTD) from HSI. There are challenges in the direct application of land-based methods to UTD, mainly because the observed target spectra, i.e., water-leaving radiance, are highly dependent on the inherent optical properties (IOPs) of water and the depth of the target, making it difficult to distinguish the target from the background. In response, research [6,10,11] may provide partial solutions. According to the implementation scene, these solutions fall into one of two categories.
In the first category, which we shall call the underwater scene, a spectral imager is attached to an autonomous underwater vehicle, which directly dives underwater to acquire HSI. A rapid target detection technique employing band selection (BS) is proposed [10]. This first uses a constrained target optimal refraction factor band selection (CTOIFBS) method to choose bands with high levels of information and low band correlation. Then, an underwater spectral imaging system incorporating the optimal subset of wavebands is built for underwater target picture capture. To detect the desired underwater targets, the CEM algorithm is finally employed. This overcomes the problems of a complex background and poor imaging environment. However, the expenses associated with underwater vehicle photography are considerable, the range of movement is limited, and BS loses a lot of spectral information while increasing its speed.
In the second category, denoted a satellite remote sensing (SRS) scene, the pipeline generally starts with IOPs estimation due to the water body’s influence on the off-water irradiance. IOPs are calculated using mathematical optimization methods (such as SOMA [12]), then a target space is constructed based on a bathymetric model to solve the issue of missing depth information, and finally, detection is carried out using a manifold learning method [11]. However, manifold learning has the drawback of being time-consuming. Deep learning has also been applied to address this issue. A network named IOPE-Net [13], in the form of an unsupervised autoencoder, was proposed to estimate IOPs and achieved a good performance. In light of this, a self-improving framework [6] is proposed, which is based on deep learning and EM theory. Firstly, anomalies are detected by a joint anomaly detector with a high threshold to obtain a guide training set, which is used to train a depth estimation network. Then, shallower pixels are screened as targets and used to train a detection network. After that, the guide training set is updated based on the detection results. Two networks are alternately iterated to boost the final detection performance. Although this framework shows a large improvement in detection performance compared to several land-based methods, it requires different iterations; thus, the issue of it being time-consuming remains, and the initial guided training set often leads to few-shot problems in practice due to its small size.
To facilitate the experiment, this paper focuses on unmanned aerial vehicle remote sensing (UAVRS) scenarios, which have the advantages of a low cost, broad applicability, high spatial resolution, and good real-time performance. In fact, in both UAVRS and SRS scenarios, underwater targets are detected from the air, so the proposed method can also be extended to SRS scenarios.
Based on the above scenarios, we propose a new framework for underwater target detection, which does not require repeated iterations or time-consuming mathematical optimization operations. We can quickly perform detection after training a model with the a priori land information of the target and the HSI of the specific water area. Deep learning methods are used throughout the pipeline. Since there is no publicly accessible HSI dataset containing underwater targets, and the received spectra are strongly influenced by the water body, which varies from water to water, it is difficult to generalize the model even if a dataset is available. Therefore, we used synthetic data to train the model for each specific water area. To solve the adaptation problem from the source domain to the target domain, we adopted domain randomization and domain adaptation strategies. Meanwhile, the spatial–spectrum process was used to utilize spatial information.
Generally speaking, the proposed framework consists of three main phrases: synthetic data, domain selection, and target detection. Firstly, we used a specific network to estimate the water quality parameters, combined the information of the target to plot the spectral distance variation curve with depth, and found the depth maximum where the spectral distance shows almost no change. Then, we selected certain target pixels, water background pixels, and depths to synthesize the underwater target observation spectra based on the bathymetric model and divided them into different domains according to depth intervals. Then, we used synthetic data and proper water background data to train the target detection network and domain prediction network. After that, we fine-tuned the detection network’s fully connected layers, i.e., the domain adaptive sub-network, according to domains. Finally, the trained model can be directly used for detection.
Aiming to quickly detect underwater targets, our main idea is to use synthetic data for training and directly apply them to real-world tasks. The closer the real-world data are to the domain of the synthetic data, the corresponding model of that domain will lead to better detection results.
In summary, the major contributions of our work are listed as follows:
  • A novel framework named TUTDF has been proposed to address the underwater detection problem based on deep learning and synthetic data theory. To the best of our knowledge, this is the first time that synthetic data have been used to train models, then directly applied to underwater target detection tasks in real environments.
  • We propose a synthetic data approach based on depth interval partitioning to solve the transfer problem from the synthetic data source domain to the real data target domain. We also develop domain prediction and domain adaptive networks to select the corresponding sourced domain model to yield more accurate detection results.
  • Due to the lack of hyperspectral underwater target datasets, we conducted extensive experiments to obtain realistic underwater target datasets. As far as we know, this is the first hyperspectral underwater target dataset with accurate depth information.
The remainder of this paper is structured as follows. Section 2 first introduces the relevant model, followed by a detailed description of the experimental procedure and some theory, before finally displaying the proposed method in more detail. The experimental results are shown in Section 3. Some of the components are discussed in Section 4. In Section 5, we summarize the entire work and draw a comprehensive conclusion.

2. Materials and Methods

In this section, a brief review of the bathymetric model of ocean hyperspectral data is given, providing a fundamental field background. The experimental implementation will be presented, which describes the data acquisition and processing in more detail. Then, some theories are introduced, and our proposed method is described in detail. Finally, the evaluation criteria, data partitioning, and other experimental details are described.

2.1. Bathymetric Model

Generally, the observed spectrum for a submerged target received by a sensor can be described as a specific equation, which is called the bathymetric model [14]. This can be formulated as Equation (1). The process described by the formula is illustrated in Figure 1. Sunlight entered the water body and was reflected by the water column or the underwater targets. The reflected rays were transmitted from the water column and received by the sensor. In addition to atmospheric attenuation, there is a large amount of light attenuation in the water, with increasing depth. The water surface may suffer from weak polarization when reflecting sunlight. For convenience, as other researchers have, we ignored the effects of the water–air interface, which can be represented by the empirical relation of Lee et al. [15]. Therefore, the target spectrum received by the sensor is affected by the following: the upper surface of the target, water quality, depth, and sunlight intensity and angle.
r ( λ ) = r ( λ ) 1 e k d ( λ ) + k u c ( λ ) H + r B ( λ ) e k d ( λ ) + k u b ( λ ) H
where λ is the wavelength, r ( λ ) is the off-water reflectance, r ( λ ) is the reflectance of the “optical deep water” [16], r B ( λ ) is the reflectance of the target to be detected as the priori information, k d ( λ ) is the downward attenuation coefficient of the light in the water, k u c ( λ ) and k u b ( λ ) are the upward attenuation coefficients of the water column and target, respectively, and H is the depth information of the underwater target.
Further, r ( λ ) , k d ( λ ) , k u c ( λ ) and k u b ( λ ) are determined by the IOPs of water: absorption coefficient a ( λ ) , backscattering coefficient b b ( λ ) . In addition, in remote sensing, k d ( λ ) is also related to the solar zenith angle θ . The commonly referred-to water properties’ estimation is the estimate a ( λ ) , b b ( λ ) from the hyperspectral image of the water surface. In this paper, we used an unsupervised water properties estimation network IOPE-Net in the form of an autoencoder to estimate a ( λ ) , b b ( λ ) . The relationship between the above parameters is as follows:
k ( λ ) = a ( λ ) + b b ( λ )
u ( λ ) = b b ( λ ) / ( a ( λ ) + b b ( λ ) )
k d ( λ ) = k ( λ ) / cos ( θ )
k u c ( λ ) = 1.03 × ( 1 + 2.4 u ( λ ) ) 0.5 k ( λ )
k u b ( λ ) = 1.04 × ( 1 + 5.4 u ( λ ) ) 0.5 k ( λ )

2.2. Experimental Implementation

Due to the novelty of hyperspectral underwater target detection, there is no publicly available dataset containing underwater targets; therefore, much of the existing work is based on simulated data, which is not a good way to check the effectiveness of the method in real environments, and applications in real environments often have more issues to consider. Therefore, we conducted extensive and effective experiments to obtain a real underwater target dataset and selected some data for validation.
In this section, we conducted an outdoor UAV experiment and an indoor pool experiment. Some preliminary conclusions were obtained from the UAV experiment. After a simplification of some influencing parameters, the indoor pool experiment was used to obtain a dataset with accurate depths.

2.2.1. Outdoor UAV Experiment

The UAV experiment was conducted in the outdoor swimming pool of Northwestern Polytechnical University (34.01°N, 108.45°E, 11:00–12:00, 17 January 2023). We first shot the underwater targets with UAV to form a basis for subsequent indoor experiments with precise depth. A brief description of the outdoor UAV experiment process and preliminary conclusions is as follows.
(1) Equipment and materials: The Gaia Sky-mini3 unmanned airborne hyperspectrometer is mainly designed for outdoor or long-range imaging tests of larger objects and has some applications that require portable operation. Gaia Sky-mini3 has a spectral range of 400–1000 nm, a spectral resolution of 5.5 nm and a spatial resolution of 1024 × 448. Then, we used DJI M600 PRO UAV, which has a vertical hovering accuracy of 0.5 m, to carry the spectrometer for the pool experiment, as shown in Figure 2a.
As exhibited in Figure 2b, we selected iron, rubber, and stone as underwater target materials of interest. Iron is the most common material for underwater vehicles, rubber is an anechoic material that is commonly used by underwater submarines to avoid sonar detection, and stone is a simulation of some underwater reefs. Detailed information on the targets is presented later in the indoor pool experiment 5.
(2) Data Acquisition: The UAV hovering vertical water surface shooting. As shown in Figure 3b, when tilting a certain angle to shoot, especially when the angle receives a direct reflection of light from the water surface, it will be beyond the water surface polarization phenomenon, affecting the quality of photography. As shown in Figure 3c, when looking at the pseudo-color image, a slight spatial distortion is clear, which is tolerable and does not affect its spectral profile. The detailed steps are as follows.
  • step1: The reference board, which is usually called ‘white board’, was taken on the ground for reflectance correction and to fix the exposure time, as shown in Figure 3a. The shooting time, latitude and longitude were recorded for solar zenith angle calculation. The ‘gray cloth’ was laid flat on the ground. The lens cap was closed to obtain the ‘black board’ data.
  • step2: The UAV flew to 20, 50, and 100 m altitudes to acquire images of the targets on the land. The ‘gray cloth’ was also taken at each altitude for atmospheric correction.
  • step3: Targets were placed at 0.89 m in the pool. Another shot was taken of the ‘white board’ on the ground and the exposure time was fixed again. The UAV flew to 20, 50, and 100 m altitudes to capture target images underwater. The ‘gray cloth’ was taken at each altitude.
  • step4: The data taken by the spectrometer were copied to the computer for processing. Using the software Spectral VIEW, atmospheric calibration was performed with “gray cloth” data, and automatic reflectance calibration was performed by selecting “whiteboard”, “blackboard” and raw data.
(3) Preliminary conclusion: Through an analysis of the obtained curves, we obtained the following preliminary conclusions and experiences. The spectral curve of the underwater target with a flat surface and uniform material does not significantly change with the change in shooting height; however, for the target with an uneven surface, with the increase in shooting height, pixel-mixing occurred and the curve of a point in the center of the surface may significantly change. An outdoor UAV needs to consider atmospheric attenuation, shooting angle, solar zenith angle, and distortion problems. Atmospheric attenuation occurred after shooting a gray cloth for atmospheric correction. The shooting angle was vertical water hovering shooting, avoiding the direct reflection of sunlight on the water’s surface. The solar zenith angle can be calculated based on latitude and longitude, as well as time.
Due to the limited depth of the pool, the water background is not ‘optically deep’, and the water quality cannot be effectively estimated. Moreover, it is difficult to obtain data for targets at different accurate depths. Therefore, we conducted the following indoor pool experiments.

2.2.2. Indoor Pool Experiment

In the anechoic pool laboratory of the College of Navigation, Northwestern Polytechnical University, there is a hydraulic lift rod, which can achieve depth variations from 0 to 3.1 m with an accuracy of 0.01 m. In addition, the pool water is deep and the indoor light is relatively weak, so the water background is optically deep. Here, we aimed to obtain underwater target data with an accurate depth.
(1) Environment and materials: The experimental site is shown in Figure 4. Atmospheric attenuation can be ignored due to the close shooting distance. There was no significant direct sunlight indoors, and unlike outdoors, here, the ‘white board’ and target were placed horizontally and shot within the same image. That is, the light incident on the ‘white board’ and the water surface was the same. We used the data from the ‘white board’ for area calibration in the software Spectral VIEW. Therefore, the solar zenith angle was not considered here and was set to 0. The hyperspectral camera still shot vertically over the water’s surface. Depth, water quality, and target material were considered. The spatial resolution for the pool experiment scene was 1.72 mm, so the pixel mixing problem can be ignored. The difference between the light source for the outdoor and the indoor pool was the intensity of the sunlight, which affects the transmission depth of light in the water body and the detectable depth. Similarly, the water quality will also affect these parameters. Although there are some differences in light intensity and water quality, the transmission mechanism of these two was consistent; they fit the same bathymetric model (1) after calibration.
The experimental targets were fixed on the fabricated bracket, and the bracket near the targets was tied with black adhesive tape to reduce the interference of reflected light. As exhibited in Figure 5, iron and rubber plates had dimensions of 1 × 0.5 m, the disc’s diameter was 0.66 m, the cylinder’s length was 0.9 m, and the diameter was 0.2 m.
(2) Data Acquisition: In Figure 6, we used a bracket to hold the target horizontally, and the bracket was connected to a hydraulic lift rod to achieve depth variations from 0 to 3.1 m with an accuracy of 0.01 m. Gaia Field–V10 is a portable push-broom spectral imager with a spectral resolution of 2.8 nm, band range of 400–1000 nm, and spatial resolution of 658 × 696. The spectral imager was positioned 4.2 m directly above the target for vertical photography, at which point the spatial resolution was 1.72 mm. The ‘white board’ with a side length of 0.15 m was placed on the water surface next to the target and included in all measurements. The reflectance calibration was performed according to Equation (7).
Ref target ( λ ) = D N target ( λ ) D N dark ( λ ) D N white ( λ ) D N dark ( λ ) Ref white ( λ )
where DN (digital number) is the original HSI pixel’s brightness value and recorded grayscale value of objects. DN target ( λ ) refers to the image brightness value of the object to be measured; DN dark ( λ ) refers to the noise generated by the spectral imager itself, which was captured by the spectral imager with the lens cover closed; DN white ( λ ) refers to the brightness value of the reference plate (commonly known as white plate); Ref white ( λ ) refers to the standard reflectance value of the ‘white board’, and a 99% reference plate was used indoors. DN target ( λ ) DN dark ( λ ) and DN white ( λ ) DN dark ( λ ) eliminated the noise in the spectral imager system. The implementation site is shown in Figure 6.
(3) Selection of Dataset Since the sunlight in the pool lab is weaker than that outdoors, we chose the depths of the human eye’s visual limit and, without a shadow, obscured a portion of the data for testing. As shown in Figure 7, they possessed 120 bands, from 400 to 780 nm. The HSI of the targets on the land was cropped at 10 × 10 pixels; water and the underwater targets had a shape of 100 × 100 after cropping. The pseudo-color pictures of the aforementioned data are as follows.

2.3. Related Theory

Many of the present artificial intelligence problems are attributed to insufficient data or available datasets being too small; even if it is relatively easy to obtain unlabeled data, the cost of manual labeling is often remarkably high. For the application of deep learning to hyperspectral underwater target detection tasks, especially low-altitude, fast detection tasks, there are unfortunately no corresponding publicly available datasets at present. Acquiring a large number of corresponding datasets is costly, which poses a significant challenge when training neural networks to perform detection tasks.
Synthetic data are a fundamental approach to solving the data problem [17], i.e., generating artificial data from scratch or using advanced data manipulation techniques to produce novel and diverse training examples, and have played a significant role in the field of computer vision and medical image processing. This paper generates synthetic data by partially “cutting” and model-based “pasting” with real data samples. Generally, the aim is to make the synthetic data as realistic as possible.
Synthetic data solve the problem of a lack of training sets. However, other problems arise. We refer to the training set generated from synthetic data as the source domain and the testing set obtained from the real world as the target domain. When we use the model trained from the source domain for target domain detection, the problem of domain transfer must be considered, i.e., how to ensure that the model from the source domain achieves good results in the target domain.
Inspired by the use of multi-path CNNs for the fine-grained recognition of cars [18] and the use of an attribute-guided feature-learning network for vehicle re-identification [19], we decided to divide multiple-source domains based on a fine-grained depth.
A multi-source domain adaptive task has multiple source domains and one target domain. The model learns from multiple source domains and applies what it has learned to predict the target domain. In general, this task seeks to identify the most useful source domain. Inspired by the domain-aware controller [20], we design a domain prediction network to select the source domain model that is to be applied to the target domain.

2.4. Proposed Method

A Transfer-Based Underwater Target Detection Framework (TUTDF) is a deep-learning-based framework for hyperspectral underwater target detection, dedicated to fast detection with known a priori information about the target on-land. We address the lack of underwater target datasets by following the bathymetric model to synthesize data to yield training sets. We used an end-to-end network that does not require a rigorous upfront noise and dimensionality reduction effort for the backbone detection network. To solve the source-to-target domain adaptation problem, we set multiple source domains according to the depth range in the synthetic data module and designed a domain prediction network to select the corresponding source domain model for detection based on the depth information. In order to utilize more spatial information, we performed a spatial–spectral process before detection to improve the performance. All the contents are shown in Figure 8. TUTDF consists of two modules: a Synthetic Dataset Module and Underwater Target Detection Module.
The Synthetic Dataset Module is the initial part of TUTDF, which is used to create the training set for the target detection module. We set the depth range and divided different source domains according to the bathymetric model to generate a large amount of training data, which were used to sample and train the subsequent network.
The Underwater Target Detection Module is the most essential part of TUTDF, which consists of the target detection network and the domain prediction network. The following describes the overall flow of this module. The HSI to be detected is input to the target detection network after the spatial–spectral process. When training the synthetic datasets, the domain prediction network can ascertain the depth of each input pixel. According to the average depth value of some of the shallowest pixels, the corresponding domain adaptive sub-network is selected and combined with the feature extraction sub-network to form a complete target detection network, which is self-sufficient as it can use an under-fitted domain prediction network to estimate the depth information. We only need an approximate depth range to determine the corresponding source domain model.

2.5. Synthetic Dataset Module

Synthetic data should resemble real data more closely and cover all potential circumstances so that samples from the target domain are included in the source domains. Given the depth H, we may use Equation (8) to yield the underwater target reflectance r s y n ( λ ) . Here, we adopted the simplified model [14], where k c c ( λ ) k u b ( λ ) . Assuming that the noise in each HSI we used is n ( λ ) , we do not need to add extra noise during the synthesizing process.
r s y n ( λ ) = r ( λ ) + n ( λ ) · 1 e k d ( λ ) + k u c ( λ ) H + r B ( λ ) + n ( λ ) · e k d ( λ ) + k u b ( λ ) H = r ( λ ) · 1 e k d ( λ ) + k u c ( λ ) H + r B ( λ ) · e k d ( λ ) + k u b ( λ ) H + n ( λ ) · 1 + e k d ( λ ) + k u b ( λ ) H e k d ( λ ) + k u c ( λ ) H r ( λ ) · 1 e k d ( λ ) + k u c ( λ ) H + r B ( λ ) · e k d ( λ ) + k u b ( λ ) H + n ( λ )
We also need to consider H. The concept of spectral distance [21] is also relevant. Spectral distance describes the difference between two spectral vectors, calculated by treating the spectral curve as a spectral vector. The spectral distance d is defined in Equation (9), which represents the difference between r B ( λ ) and r s y n ( λ ) .
d = r s y n ( λ ) r B ( λ ) 2
As shown in Figure 9, the spectral distance increases with depth, and the growth rate gradually decreases to zero. After reaching a particular depth, r s y n ( λ ) becomes the optical deep water r ( λ ) , and the influence of the spectral imager resolution and noise renders the observed target spectrum indistinguishable from the water background. Therefore, the spectral distance curves of different materials were constructed under the specific water condition, with a maximum depth of H d e e p .
A brief argument for the validity of using synthetic data is presented as follows. Firstly, the synthetic data are based on the model, which is reproducible and stable. Secondly, the spectral distance curves of the synthetic data match well with the real situation in Figure 9. Finally, the results regarding the real data in Section 4 will be more effective at proving the validity of the synthetic data.
As there are still differences between the real and synthetic spectra, they cannot be completely identical at all depths. As a result, it is expected that the distribution in the synthetic domain encompasses the real distribution. To this end, the depth interval [ 0 , H d e e p ] was divided into N overlapping sub-intervals, with the length of the interval being set to gradually increase from shallow to deep, as shown in Figure 10.
Each depth interval has M depths taken at equal intervals, and the data synthesized in different intervals belong to different domains D d , d = 1 , 2 , 3 , , N . Figure 11 illustrates the whole process of synthesizing data, which is performed in the following steps.
  • step1: Train a water quality estimation model f I O P ( ) unsupervised with IOPE-Net from HSI of water areas.
  • step2: Select P pixels from target donates r B i , i = 1 , 2 , 3 , , P and Q pixels from water donates r j , j = 1 , 2 , 3 , , Q . Calculate the water quality parameters a j ( λ ) , b b j ( λ ) = f I O P ( r j ( λ ) ) .
  • step3: Use r B i ( λ ) ¯ and r j ( λ ) ¯ to plot the spectral distance, combine the resolution conditions, set the maximum depth H d e e p , divide into N overlapping sub-intervals, and select M depths at each sub-interval.
  • step4: According to Equation (8), P × Q × M underwater target samples are generated at each interval, and classified in domain D d .
  • step5: The pixels in domain D d and an equal number of optically deep water pixels are used as the training set for the domain adaptive sub-network. The data of N domains are used as the training set for the domain prediction network.
Figure 12 shows the results of partial synthesis data. In this case, we separately selected one pixel from iron and water. The reflectance decreases with increasing depth until r ( λ ) , and there is an overlap between adjacent domains.

2.6. Target Detection Module

The module consists of two networks, the target detection network and the domain prediction network, described below.

2.6.1. Domain Predication Network

As can be seen from Figure 13, the first half of the network is a depth estimation network. During the training phase, the network employs unsupervised training in the form of an autoencoder. The input pixel is denoted by x . Using a two-layer 1D-CNN encoder, the estimated depth h ^ is obtained. Then, we reconstructed the pixel x ^ using h ^ and the bathymetric model in the decoder part. Finally, the network parameters were updated using the back-propagation of the loss function L.
L = L M S E + λ S L S + λ H H 2
L M S E = x x ^ 2
L S = 1 π a r c c o s x · x ^ x 2 x ^ 2
where · 2 is the l 2 norm of a given vector. L M S E is the mean square error loss function, describing the euclidean distance between x and x ^ . L S is the spectral angle loss function, which describes the similarity in shape between x and x ^ . H 2 is the depth-constraint loss function, which is a very important term in the loss function that keeps the estimated depth from being too large. λ S and λ H are the corresponding weighting coefficients of the loss function L M S E and H 2 .
Since the depth estimate of the target is always small, we selected the smallest part of the depth value as the overall target depth. During the detection phase, the to-be-detected HSI is first fed into the depth estimation network to select the minimum S depths with a depth value of h ( s ) ; then, the depth values are processed by the domain prediction operation to select the model used in the domain adaptive network. The domain prediction operation is as follows.
h ¯ = s = 1 S h ( s ) S
The value of S was determined by considering the size of the target in the camera’s field of view. If the field of view is large and the target is small, the selected S is larger than the number of target points, which means that the estimated target depth will be large; thus, an inappropriate domain will be selected. If S is set too small, it may be disturbed by individual anomalies in the water background and estimate an inappropriate approximate overall target depth. Therefore, the setting of S can be expressed as follows.
S = N f i l e d S t a r g e t S f i l e d
where N f i l e d is the number of spatial pixels in the hyperspectral image captured by the camera, S t a r g e t is the area determined by the target being detected (set to 1/4 of the area of the upper surface of the target being detected, as only a part of the target is captured), and S f i l e d is the area of the field of view at a fixed shooting height.
For depth intervals of different domains, this setting ensures that no certain depth will occur at the boundary of the selected field.
D d = D 1 , 0 h ¯ < H 12 + H 21 / 2 D 2 , H 12 + H 21 / 2 h ¯ < H 22 + H 31 / 2 D 2 , H 12 + H 21 / 2 h ¯ < H 22 + H 31 / 2 D N , H ( N 1 ) 2 + H N 1 / 2 h ¯ < N

2.6.2. Target Detection Network

The network input is a pixel, i.e., a hyperspectral curve, which is regarded as a sequence input, and the output is the probability of the target to be detected. Two sub-networks comprise the target detection network: feature extraction sub-network and domain adaptive sub-network, as seen in Figure 14.
The feature extraction network is a generic backbone network. Recently, a specific deep network [22] successfully applied to long sequence data (ambulatory electrocardiograms) detection and classification. Inspired by that, we use a fully convolutional deep neural network as the backbone feature extraction network to achieve high performance without extra preprocessing (noise reduction, dimensionality reduction). The network simply inputs the original spectral sequence and generates the feature map through multilayer 1-D CNN with residual connectivity.
With a pixel x = x i i = 1 B containing B bands, for the feature extraction network, the output of the first block is:
y ( 1 ) = h w ( 1 ) x + b ( 1 )
the output of the second block is:
y ( 2 , 1 ) = h w ( 2 , 1 ) y ( 1 ) + b ( 2 , 1 ) y ( 2 , 2 ) = w ( 2 , 2 ) y ( 2 , 1 ) + b ( 2 , 2 ) y ( 2 ) = y ( 2 , 2 ) y ( 1 )
and the outputs of the successor blocks are:
y ( t , 1 ) = w ( t , 1 ) h y ( t 1 ) + b ( t , 1 ) , y ( t , 2 ) = w ( 3 , 2 ) h y ( t , 1 ) + b ( t , 2 ) , t 3 y ( t ) = y ( t , 2 ) y ( t 1 ) ,
where w ( t , 1 ) , w ( t , 2 ) and b ( t , 1 ) , b ( t , 2 ) denote the first and second convolutional layer weight and bias parameters of the t t h block, respectively; ∗ represents the convolution process, h is the nonlinear activation function, and ⊕ denotes the residual connection. To effectively solve the gradient disappearance problem, the ReLU [23] activation function is utilized; batch normalization and dropout are applied to accelerate network convergence. The residual connection is used to alleviate the problem of network degradation caused by gradient disappearance or explosion, and redundant information is removed by max pooling when the residual connection is formed. After every two residual blocks, the network channels are reduced by half. When the remaining connections do not add up to the same dimension, zero-padding is performed for the residual channel dimension. The domain adaptive network is a fully connected network. The softmax function obtains the probabilities that the curve belongs to the target and the water background. The network parameters are updated using cross-entropy loss function backpropagation.
Fine-tuning can train a model quickly, with a relatively small amount of data, and still achieve good results. A target detection network consists of a feature extraction sub-network and domain adaptive sub-network. To quickly train the models, we trained the entire network with the first domain’s dataset, then froze the feature extraction sub-network parameters and retrained the entire network with remainder datasets separately. In this way, we can obtain the corresponding domain adaptive sub-network model parameters.
When detected in the real scene, the spatial–spectral process is performed before the input network. Spatial consistency is a characteristic of feature distribution in hyperspectral images, i.e., features of the same category are concentrated in the actual environment, and the smaller the distance between features, the greater the probability that they belong to the same category.
For the pixel x ( m , n ) = x i ( m , n ) i = 1 B with spatial location ( m , n ) in HSI, the region of the size ω × ω at the center of the pixel but not containing the pixel itself is called the nearest neighbor space of x ( m , n ) , where ω ω ω = 2 k + 1 , k Z + . Usually, ω = is taken as a positive odd number to indicate the size of the region so that it has a center. Additionally, ω can be considered as the filter kernel size or the perceptual field size.
Ω x ( m , n ) = x ( p , q ) p [ m a , m + a ] , q [ n a , n + a ]
where a = ( ω 1 ) / 2 if there is a vacancy in the near-neighbor space of the pixel, such as at the corner or edge; therefore, this type of pixel can fill the vacant position in the near-neighbor space. Figure 15 shows the nearest neighbor space when ω = 3 , and the blue area is Ω ( x ( m , n ) ) .
According to the principle of spatial consistency, the central pixel x ( m , n ) is replaced by the weighted sum of the pixels in the near-neighbor space; since each pixel contains spectral information, this processing combines the spatial and spectral information of the image, which can effectively reduce the phenomenon of foreign objects with the same spectrum and eliminate some anomalies. The recalculated pixel x ^ ( m , n ) can be expressed as:
x ^ ( m , n ) = w ( p , q ) x ( p , q ) w ( p , q ) , x ( p , q ) Ω x ( m , n )
where w ( p , q ) = e x ( m , n ) x ( p , q ) 2 / γ is the weight of any pixel in the near-neighbor space. γ is calculated as the square of the average distance between the central pixel and any other pixel in its near-neighbor space, excluding itself. The physical meaning of the size of w ( p , q ) is that the more similar the pixel features, the greater the degree of interaction between them, and the greater the differences between the features, the smaller the degree of interaction between them.
In our work, since the water background and the targets are relatively simple, for simplicity, we set ω = 3 and used nearest neighbor spatial averaging to replace the center pixel.

2.7. Experimental Details

(1) Evaluation Criteria: The traditional receiver operating characteristic (ROC) curve depicts the correlation between the false alarm rate (FAR) P F and target detection probability P D , which can be formulated as Equations (21) and (22). The P F represents the portion of falsely detected pixels for the entire image:
P F = N F N
where N F stands for the number of falsely detected pixels and N refers to the total number of pixels in the image. The target detection probability P D denotes the ratio between correctly detected pixels and target pixels among the whole image:
P D = N C N T
where N C is the amount of correctly detected pixels and N T refers to the number of target pixels in the entire image. Once the detection result maps are given, we can set a series threshold τ to obtain many ( P D , P F ) to generate the ROC curve. The closer the ROC curve of ( P D , P F ) is to the upper left corner, the better the detection effect. Meanwhile, the area under the curve (AUC) of ( P D , P F ) is used for quantitative evaluation. The larger the AUC value, the better the detection effect.
However, a traditional ROC curve of ( P D , P F ) can only be used to evaluate the effectiveness of a detector, not target detectability (TD) or background suppressibility (BS). For a more comprehensive evaluation, we adopted a new 3D ROC curve [24,25]. Compared with the traditional ROC curve, the 3D ROC curve criterion, in the form of its three derived 2D ROC curves of ( P D , P F ), ( P D , τ ), and ( P F , τ ) can further be used to evaluate the effectiveness of a detector and its TD and BS, which can be quantified as:
0 A U C T D = A U C ( D , F ) + A U C ( D , τ ) 2
1 A U C B S = A U C ( D , F ) A U C ( F , τ ) 2
where A U C ( D , F ) , A U C ( D , τ ) and A U C ( F , τ ) represent the AUC value of the ROC curve of ( P D , P F ), ( P D , τ ) and ( P F , τ ), respectively. The calculated A U C T D and A U C B S are employed to quantitatively characterize the detector’s TD and BS, respectively.
(2) Experimental Settings: We used the spectral imager with SpecView software to calibrate the reflectance of the raw images and utilized ENVI 5.3 to perform HSI cropping and waveband selection. Subsequent experiments were based on an Intel(R) Core(TM) i9-12900KF CPU, GeForce RTX 3080 Ti GPU, 64 GB RAM, Windows 11 OS, and the code was based on Python 3.9 and the deep learning framework PyTorch 1.12.1. The relevant parameters of the experimental dataset are shown in Table 1, Table 2 and Table 3.

3. Results

To demonstrate the superiority of our proposed method, we compared TUTDF with traditional land-based hyperspectral target detection methods—MF [26], CEM, ACE [27], OSP, SAM [28], and TCIMF [29]—using different materials and depths. The ground truth of the datasets in Figure 16, Figures 18 and 19 are manual tagging according to pseudo-color images.

3.1. Iron Plate

First and foremost, we will demonstrate the detection performance from a visual standpoint by Figure 16. Depending on the visual judgment, the TUTDF possessed the slightest visible difference from the ground truth. Furthermore, as compared to other detectors, TUTDF was able to reduce the negative influence of background pixels by detecting the fewest background pixels. However, traditional land-based target detection algorithms fared poorly across all datasets because the target and background were no longer independent in underwater contexts.
To conduct a qualitative analysis of the detection results, the ROC curves of ( P D , P F ), ( P D , τ ) and ( P F , τ ) were plotted in Figure 17 as a further three critical comparison metrics. The ROC curves of ( P D , P F ) of our method remain over other curves in all depths. Among the other detectors, OSP and SAM perform well when the target is shallower, but when the target is deeper, their performance decreases. The ACE detector fails at all depths. For the ROC curves of ( P D , τ ), ACE occurred at the lower right, which demonstrated its invalid TD. TUTDF was located upwards from the other methods used for comparison at each depth, except 1.6 m, where it was placed slightly lower than some other methods. Moreover, the ROC curves of ( P F , τ ) of the TUTDF were located at the bottoms of other methods used for comparison at each depth, except ACE. However, ACE is invalid. Based on these three curves, we could conclude that TUTDF has a better detection performance compared with the lower FARs, even when the target is deeper, and other methods have failed.
For a quantitative comparison, as shown in Table 4, the AUC values of ( P D , P F ), ( P D , τ ), ( P F , τ ), TD and BS were also calculated. The AUC(D,F) values of TUTDF were higher than other detectors and remained at 1.00, while other detectors’ decreased with depth. Although the AUC ( P D , τ ) value of our proposed method at 1.6 m is slightly lower than OSP, when it comes to a more accurate TD-evaluating indicator AUC T D , TUTDF is at least 0.16 higher than other detectors, proving its excellent target detection abilities. Furthermore, TUTDF attains the lowest FAR among all valid detectors, and the corresponding AUC ( P F , τ ) values were under 0.17 at all depths. Moreover, in terms of BS, TUTDF is also better than other detectors. When the target is at a deeper position, the AUC B S values of TUTDF were much higher than other detectors, which can effectively suppress the interference of the water background.

3.2. Stone Disc

As shown in Figure 18 and Table 5, the test performance is consistent with iron when performed on the stone discs. From a material standpoint, the detection performance remains superior, even though the stone material has higher intraclass variability than the iron. Accordingly, we can conclude that the deep learning method trained using synthetic data can learn the underwater information of target pixels of the same material with differences and detect the target with low FAR.

3.3. Rubber Plate

The results of the rubber plate detection are shown in Figure 19 and Table 6. Due to the strong absorption properties of rubber material, the detection performance of the rubber plate is lower compared to iron and stone and can only be detected within a shallow depth range. However, our method still performed the best in terms of TD and BS. This further confirms the generalization ability of our proposed method, which can detect targets of different materials. The result also indicates that rubber, which is a typical submarine anechoic material used to avoid sonar detection, is also effective in avoiding hyperspectral detection in the visible band. Therefore, to more effectively detect underwater submarines, it is often necessary to use infrared bands or detect heat trails.
Due to the limits of the spectral imager’s shooting time and the processing of massive data volumes, achieving a real-time performance is difficult. We are primarily concerned with the time required to make a complete detection. All detection methods were conducted on CPU, and the final results were the average of the five run-times. Table 7 shows that the execution time of TUTDF is much lower than the other methods used in our datasets (spatial size of 100 × 100, 120 bands). As seen in Table 8, the most time-consuming step is the domain prediction, subsequent parallel optimization with GPU acceleration may further reduce the TUTDF execution time, which is not discussed in detail here.
In sum, TUTDF can attain a superior detection performance in TD and BS and more in-depth detection. Additionally, it can quickly perform underwater target detection. Considering all aspects of the detection results, we are confident in concluding that the proposed detection framework, trained using synthetic data, is effective, reliable, and capable of achieving a remarkable detection performance in most underwater materials.

4. Discussion

This section discusses and analyzes the effectiveness of the domain adaptive method and the spatial–spectral process.

4.1. Effectiveness Evaluation of the Domain Adaptation

To evaluate the effectiveness of the domain adaptive method, we conducted a performance analysis of the domain adaptive method without the spatial–spectral process. Since the shallower targets all have higher A U C ( D , F ) values, we chose the deeper targets for comparison to reflect the differences more prominently, and conducted experiments on 1.6 m iron plates and 1.6 m stone discs, respectively. This process followed the proposed pipeline. The method without domain adaptation aims to sample 100 depths at equal intervals from 0 to H d e e p ; then, the network was trained and could detect the target directly.
The results under the conditions of a 1.6 m iron plate and stone disc are shown in Figure 20 and Table 9. Qualitatively, taking the domain adaptive method makes the detection results more clearly distinguishable; quantitatively, using the domain adaptive method makes the A U C ( D , F ) values significantly improve by 0.2.
Table 10 and Table 11 show the interval settings and depth estimation results. Suitable source domains can be selected without strictly accurate depth estimation results, and the corresponding domain-adaptive sub-networks can be used to perform the detection task more efficiently. Here, the parameter S in Equation (13) was set to 100.

4.2. Effectiveness Evaluation of the Spatial–Spectral Process

Unfortunately, noise-free HSI are not available on real data to evaluate the noise reduction performance. However, the effects of HSI before and after processing on the subsequent detection performance are considered as an indirect objective evaluation criterion. To discuss the influence of the spatial–spectral process on the detection performance, the table below compares the detection performance of submerged iron, stone, and rubber.
Intuitively, in Table 12, Table 13 and Table 14, since the spatial–spectral process contains an approximate mean filtering of the images, the detection results are less noisy and the target was more easily distinguishable from the background after the procedure was executed. After processing, the A U C ( D , F ) value of the detection results is qualitatively enhanced, particularly at greater depths.
The spatial–spectral process can enhance the efficiency of hyperspectral underwater target detection, especially when the target depth is relatively deeper and the distinction between the target and background is low. As seen in Table 8, this operation does not significantly increase detection time and is very effective for spatial information usage.

5. Conclusions

This paper used a deep learning method to detect underwater targets from HSI. For a low-altitude remote sensing scene, a framework named TUTDF is proposed. This particular framework consists of two large modules: the synthetic data module and the target detection module. Due to the difficulty of obtaining hyperspectral underwater target datasets, we proposed a transfer-based approach, which trained the synthetic data-based networks, then transferred them to real data scenarios. The proposed framework first developed an unsupervised network to obtain water-quality parameters, combined with the a priori information of the target that is to be detected and set different depth intervals to synthesize underwater target datasets for multiple source domains to train the networks. Then, we designed a domain prediction network to select the source domain that is closest to the target domain based on the depth information. Finally, we can transfer the trained networks to a real scenario to directly perform the detection task.
To verify the effectiveness of the method in real scenarios, we conducted extensive and quantitative indoor and outdoor experiments and then obtained hyperspectral underwater target datasets with accurate depth labels. The detection results for real data with different materials and depths show that the framework trained using synthetic data is valid and reliable, and the transfer from synthetic data domains to real-world domains is feasible. The proposed framework has a definite advantage in terms of TD, BS, and detection speed over traditional land-based detection methods.
In addition, we carried out some work in UAV underwater data acquisition; focusing on how to optimize the data acquisition and calibration methods to achieve real-time UAV underwater target detection will be the subject of future work.

Author Contributions

Conceptualization, Z.L.; methodology, Z.L.; software, Z.L and P.Z.; validation, Z.L., P.Z., and Q.L.; formal analysis, Z.L., Y.S.; investigation, Z.L., T.L., Y.S. and L.Z.; resources, J.L.; data curation, Z.L., P.Z., L.Z. and X.L.; writing—original draft preparation, Z.L.; writing—review and editing, Z.L., Q.L., and P.Z.; visualization, Z.L, P.Z., and Q.L.; supervision, J.L.; project administration, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jing, W.; Chang, C.I. Independent component analysis-based dimensionality reduction with applications in hyperspectral image analysis. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1586–1600. [Google Scholar] [CrossRef]
  2. Chang, C.I. Hyperspectral Imaging: Techniques for Spectral Detection and Classification; Springer: New York, NY, USA, 2003; Available online: http://www.springer.com/cda/content/document/productFlyer/productFlyer-CN_978-0-306-47483-5.pdf?SGWID=0-0-1297-33380812-bookseller (accessed on 27 December 2022).
  3. Chang, C.I. Hyperspectral Target Detection: Hypothesis Testing, Signal-to-Noise Ratio, and Spectral Angle Theories. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–23. [Google Scholar] [CrossRef]
  4. Chang, C.I. Statistical Detection Theory Approach to Hyperspectral Image Classification. IEEE Trans. Geoence Remote Sens. 2018, 57, 2057–2074. [Google Scholar] [CrossRef]
  5. Nie, J.; Guo, J.; Xu, Q. Object-Level Hyperspectral Target Detection Based on Spectral-Spatial Features Integrated YOLOv4-Tiny Network. In Proceedings of the 2022 4th International Conference on Image, Video and Signal Processing, IVSP 2022, Singapore, 18–20 March 2022. [Google Scholar]
  6. Qi, J.; Wan, P.; Gong, Z.; Xue, W.; Yao, A.; Liu, X.; Zhong, P. A Self-Improving Framework for Joint Depth Estimation and Underwater Target Detection from Hyperspectral Imagery. Remote Sens. 2021, 13, 1721. [Google Scholar] [CrossRef]
  7. Reed, I.; Yu, X. Adaptive multiple-band CFAR detection of an optical pattern with unknown spectral distribution. IEEE Trans. Acoust. Speech Signal Process. 1990, 38, 1760–1770. [Google Scholar] [CrossRef]
  8. Harsanyi, J.C. Detection and Classification of Subpixel Spectral Signatures in Hyperspectral Image Sequences. Ph.D. Thesis, University of Maryland Baltimore County, Baltimore, MD, USA, 1993. Available online: https://www.proquest.com/openview/03a974e5a3ca73bd78f5b5d1e5d1e887/1?pq-origsite=gscholar&cbl=18750&diss=y (accessed on 27 December 2022).
  9. Harsanyi, J.; Chang, C.I. Hyperspectral image classification and dimensionality reduction: An orthogonal subspace projection approach. IEEE Trans. Geosci. Remote Sens. 1994, 32, 779–785. [Google Scholar] [CrossRef]
  10. Fu, X.; Shang, X.; Sun, X.; Yu, H.; Song, M.; Chang, C.I. Underwater hyperspectral target detection with band selection. Remote Sens. 2020, 12, 1056. [Google Scholar] [CrossRef]
  11. Gillis, D. Detection of underwater objects in hyperspectral imagery. In Proceedings of the 2016 8th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing (WHISPERS), Los Angeles, CA, USA, 21–24 August 2016. [Google Scholar]
  12. Gillis, D.B.; Bowles, J.H.; Montes, M.J.; Miller, W.D. Deriving bathymetry and water properties from hyperspectral imagery by spectral matching using a full radiative transfer model. Remote Sens. Lett. 2020, 11, 903–912. [Google Scholar] [CrossRef]
  13. Qi, J.; Xue, W.; Gong, Z.; Zhang, S.; Yao, A.; Zhong, P. Hybrid Sequence Networks for Unsupervised Water Properties Estimation From Hyperspectral Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 3830–3845. [Google Scholar] [CrossRef]
  14. Albert, A.; Mobley, C.D. An analytical model for subsurface irradiance and remote sensing reflectance in deep and shallow case-2 waters. Opt. Express 2003, 11, 2873–2890. [Google Scholar] [CrossRef]
  15. Lee, Z.; Carder, K.L.; Mobley, C.D.; Steward, R.G.; Patch, J.S. Hyperspectral remote sensing for shallow waters. I. A semianalytical model. Appl. Opt. 1998, 27, 6329–6338. [Google Scholar] [CrossRef] [PubMed]
  16. Gillis, D.B. An underwater target detection framework for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1798–1810. [Google Scholar] [CrossRef]
  17. Nikolenko, S.I. Synthetic Data for Deep Learning; Springer: Cham, Switzerlad, 2021; Volume 174. [Google Scholar]
  18. Wang, H.; Peng, J.; Zhao, Y.; Fu, X. Multi-Path Deep CNNs for Fine-Grained Car Recognition. IEEE Trans. Veh. Technol. 2020, 69, 10484–10493. [Google Scholar] [CrossRef]
  19. Wang, H.; Peng, J.; Chen, D.; Jiang, G.; Zhao, T.; Fu, X. Attribute-Guided Feature Learning Network for Vehicle Reidentification. IEEE MultiMedia 2020, 27, 112–121. [Google Scholar] [CrossRef]
  20. Hu, S.; Liao, Z.; Zhang, J.; Xia, Y. Domain and content adaptive convolution for domain generalization in medical image segmentation. arXiv 2021, arXiv:2109.05676. [Google Scholar]
  21. Deborah, H.; Richard, N.; Hardeberg, J.Y. A comprehensive evaluation of spectral distance functions and metrics for hyperspectral image processing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3224–3234. [Google Scholar] [CrossRef]
  22. Hannun, A.Y.; Rajpurkar, P.; Haghpanahi, M.; Tison, G.H.; Bourn, C.; Turakhia, M.P.; Ng, A.Y. Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network. Nat. Med. 2019, 25, 65–69. [Google Scholar] [CrossRef]
  23. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  24. Chang, C.I. An Effective Evaluation Tool for Hyperspectral Target Detection: 3D Receiver Operating Characteristic Curve Analysis. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5131–5153. [Google Scholar] [CrossRef]
  25. Chang, C.I. Comprehensive Analysis of Receiver Operating Characteristic (ROC) Curves for Hyperspectral Anomaly Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–24. [Google Scholar] [CrossRef]
  26. Manolakis, D.G.; Shaw, G.A. Directionally constrained or constrained energy minimization adaptive matched filter: Theory and practice. In Proceedings of the Imaging Spectrometry VII, San Diego, CA, USA, 17 January 2002; Volume 4480. [Google Scholar]
  27. Manolakis, D.; Shaw, G. Detection algorithms for hyperspectral imaging applications. IEEE Signal Process. Mag. 2002, 19, 29–43. [Google Scholar] [CrossRef]
  28. Kruse, F.A.; Lefkoff, A.; Boardman, J.; Heidebrecht, K.; Shapiro, A.; Barloon, P.; Goetz, A. The spectral image processing system (SIPS)—Interactive visualization and analysis of imaging spectrometer data. Remote Sens. Environ. 1993, 44, 145–163. [Google Scholar] [CrossRef]
  29. Ren, H.; Chang, C.I. Target-constrained interference-minimized approach to subpixel target detection for hyperspectral images. Opt. Eng. 2000, 39, 3138–3145. [Google Scholar] [CrossRef]
Figure 1. A depiction of the transmission process.
Figure 1. A depiction of the transmission process.
Remotesensing 15 01023 g001
Figure 2. Photos of the experiment site. (a) UAV carrying Gaia Sky-mini3 hovering above the pool. (b) Targets placed at 0.89 m underwater.
Figure 2. Photos of the experiment site. (a) UAV carrying Gaia Sky-mini3 hovering above the pool. (b) Targets placed at 0.89 m underwater.
Remotesensing 15 01023 g002
Figure 3. Photos of the experiment site. (a) Capture ‘white board’. (b) Water surface polarization and reflection. (c) The UAV-captured image with slight spatial distortion.
Figure 3. Photos of the experiment site. (a) Capture ‘white board’. (b) Water surface polarization and reflection. (c) The UAV-captured image with slight spatial distortion.
Remotesensing 15 01023 g003
Figure 4. Anechoic pool site. The crane can be moved horizontally for the easy loading and unloading of targets. The lift rod can move up and down to place targets at different depths underwater.
Figure 4. Anechoic pool site. The crane can be moved horizontally for the easy loading and unloading of targets. The lift rod can move up and down to place targets at different depths underwater.
Remotesensing 15 01023 g004
Figure 5. Photos of the targets and reference plate on the water. (a) Iron plate. (b) Rubber plate. (c) Stone disc. (d) Iron column. (e) Rubber column. (f) Reference plate.
Figure 5. Photos of the targets and reference plate on the water. (a) Iron plate. (b) Rubber plate. (c) Stone disc. (d) Iron column. (e) Rubber column. (f) Reference plate.
Remotesensing 15 01023 g005
Figure 6. The spectra collection configurations. (a) The experimental site. Note that a halogen lamp is used for brightening, but the data in this paper did not use this. (b) The field of view of the spectral imager.
Figure 6. The spectra collection configurations. (a) The experimental site. Note that a halogen lamp is used for brightening, but the data in this paper did not use this. (b) The field of view of the spectral imager.
Remotesensing 15 01023 g006
Figure 7. The selected dataset includes iron, stone and rubber at different depths.
Figure 7. The selected dataset includes iron, stone and rubber at different depths.
Remotesensing 15 01023 g007
Figure 8. The general framework of our proposed method.
Figure 8. The general framework of our proposed method.
Remotesensing 15 01023 g008
Figure 9. The spectral distance of iron and stone on synthetic data and real data. (a) Iron. (b) Stone.
Figure 9. The spectral distance of iron and stone on synthetic data and real data. (a) Iron. (b) Stone.
Remotesensing 15 01023 g009
Figure 10. Description of the depth interval division.
Figure 10. Description of the depth interval division.
Remotesensing 15 01023 g010
Figure 11. The synthetic dataset module.
Figure 11. The synthetic dataset module.
Remotesensing 15 01023 g011
Figure 12. Part of the synthetic data curve.
Figure 12. Part of the synthetic data curve.
Remotesensing 15 01023 g012
Figure 13. The domain prediction network.
Figure 13. The domain prediction network.
Remotesensing 15 01023 g013
Figure 14. Target detection network.
Figure 14. Target detection network.
Remotesensing 15 01023 g014
Figure 15. The representation of near-neighbor space.
Figure 15. The representation of near-neighbor space.
Remotesensing 15 01023 g015
Figure 16. (a) Ground truth of the 0.1, 0.5, 1.0, 1.6 m submerged iron plate. (b) TUTDF. (c) MF. (d) CEM. (e) ACE. (f) OSP. (g) SAM. (h) TCIMF.
Figure 16. (a) Ground truth of the 0.1, 0.5, 1.0, 1.6 m submerged iron plate. (b) TUTDF. (c) MF. (d) CEM. (e) ACE. (f) OSP. (g) SAM. (h) TCIMF.
Remotesensing 15 01023 g016
Figure 17. The ROC curves of ( P D , P F ), ( P D , T h r e s h o l d ), ( P F , T h r e s h o l d ) for the compared underwater target detection methods on IRON at different depths (0.1, 0.5, 1.0, 1.6 m from top to bottom).
Figure 17. The ROC curves of ( P D , P F ), ( P D , T h r e s h o l d ), ( P F , T h r e s h o l d ) for the compared underwater target detection methods on IRON at different depths (0.1, 0.5, 1.0, 1.6 m from top to bottom).
Remotesensing 15 01023 g017
Figure 18. (a) Ground truth of the 0.4, 0.7, 1.6, 1.9 m submerged stone disc. (b) TUTDF. (c) MF. (d) CEM. (e) ACE. (f) OSP. (g) SAM. (h) TCIMF.
Figure 18. (a) Ground truth of the 0.4, 0.7, 1.6, 1.9 m submerged stone disc. (b) TUTDF. (c) MF. (d) CEM. (e) ACE. (f) OSP. (g) SAM. (h) TCIMF.
Remotesensing 15 01023 g018
Figure 19. (a) Ground truth of the 0.1, 0.3, 0.6, 0.9 m submerged rubber plate. (b) TUTDF. (c) MF. (d) CEM. (e) ACE. (f) OSP. (g) SAM. (h) TCIMF.
Figure 19. (a) Ground truth of the 0.1, 0.3, 0.6, 0.9 m submerged rubber plate. (b) TUTDF. (c) MF. (d) CEM. (e) ACE. (f) OSP. (g) SAM. (h) TCIMF.
Remotesensing 15 01023 g019
Figure 20. The result of the methods with and without a domain adaptive for a 1.6 m iron plate, stone disc.
Figure 20. The result of the methods with and without a domain adaptive for a 1.6 m iron plate, stone disc.
Remotesensing 15 01023 g020
Table 1. Target detection network training dataset parameters.
Table 1. Target detection network training dataset parameters.
MaterialTarget PixelsWater PixelsNumber of DepthSynthetic DataWater DataSource Domain
Iron101010010,00010,0006
Stone101010010,00010,0006
Rubber101010010,00010,0005
Table 2. Depth estimation network training dataset parameters.
Table 2. Depth estimation network training dataset parameters.
MaterialTarget PixelWater PixelDepth RangeNumber of Depth
Iron10100–2.5 m250
Stone10100–2.0 m200
Rubber10100–1.0 m100
Table 3. Test dataset parameters.
Table 3. Test dataset parameters.
MaterialSizeDepth
Iron100 × 1000.1 m, 0.5 m, 1.0 m, 1.6 m
Stone100 × 1000.4 m, 0.7 m, 1.6 m, 1.9 m
Rubber100 × 1000.1 m, 0.3 m, 0.6 m, 0.9 m
Table 4. AUC values calculated from the three 2D ROC curves of different detection results on iron at various depths.
Table 4. AUC values calculated from the three 2D ROC curves of different detection results on iron at various depths.
DepthDetectorAUC( D , F )AUC( D , τ )AUC( F , τ )AUC TD AUC BS
0.1 mTUTDF1.000.750.031.750.97
OSP1.000.500.501.500.50
SAM1.000.050.051.050.95
ACE0.490.500.50.99−0.01
CEM0.590.590.561.180.03
MF0.590.500.501.090.09
TCIMF0.610.500.501.110.11
0.5 mTUTDF1.000.670.031.670.97
OSP1.000.270.271.270.73
SAM0.900.390.251.290.65
ACE0.580.030.020.610.56
CEM0.520.650.641.17−0.12
MF0.590.640.641.23−0.05
TCIMF0.500.590.581.09−0.08
1.0 mTUTDF1.000.690.111.690.89
OSP0.990.380.381.370.61
SAM0.830.350.241.180.59
ACE0.570.070.050.640.52
CEM0.540.520.501.060.04
MF0.510.510.501.020.01
TCIMF0.520.530.511.050.01
1.6 mTUTDF1.000.540.171.540.83
OSP0.780.600.481.380.30
SAM0.620.280.250.900.37
ACE0.540.070.060.610.48
CEM0.600.560.511.160.09
MF0.590.550.511.140.08
TCIMF0.510.480.480.990.03
The bold entries represent the best performance in each column at a certain depth.
Table 5. AUC values calculated from the three 2D ROC curves of different detection results on STONE at various depths.
Table 5. AUC values calculated from the three 2D ROC curves of different detection results on STONE at various depths.
DepthDetectorAUC( D , F )AUC( D , τ )AUC( F , τ )AUC TD AUC BS
0.4 mTUTDF1.000.750.031.750.97
OSP1.000.500.501.500.50
SAM1.000.050.051.050.95
ACE0.530.500.501.030.03
CEM0.510.590.561.10−0.05
MF0.510.500.501.010.01
TCIMF0.440.500.500.94−0.06
0.7 mTUTDF1.000.670.031.670.97
OSP10.270.271.270.73
SAM1.000.390.251.390.75
ACE0.530.030.020.560.51
CEM0.430.650.641.08−0.21
MF0.470.640.641.11−0.17
TCIMF0.430.590.581.02−0.15
1.6 mTUTDF0.980.690.111.670.87
OSP0.910.380.381.290.53
SAM0.850.350.241.200.61
ACE0.510.070.050.580.46
CEM0.500.520.51.020.00
MF0.490.510.501.00−0.01
TCIMF0.380.530.510.91−0.13
1.9 mTUTDF0.810.540.171.350.64
OSP0.770.600.481.370.29
SAM0.730.280.251.010.48
ACE0.470.070.060.540.41
CEM0.490.560.511.05−0.02
MF0.490.550.511.04−0.02
TCIMF0.320.480.480.8−0.16
The bold entries represent the best performance in each column at a certain depth.
Table 6. AUC values calculated from the three 2D ROC curves of different detection results on RUBBER at various depths.
Table 6. AUC values calculated from the three 2D ROC curves of different detection results on RUBBER at various depths.
DepthDetectorAUC( D , F )AUC( D , τ )AUC( F , τ )AUC TD AUC BS
0.1 mTUTDF0.990.750.031.750.97
OSP0.990.500.501.500.50
SAM0.990.050.051.050.95
ACE0.530.500.500.99−0.01
CEM0.690.590.561.180.03
MF0.700.500.501.090.09
TCIMF0.730.500.501.110.11
0.3 mTUTDF0.980.670.031.670.97
OSP0.850.270.271.270.73
SAM0.710.390.251.290.65
ACE0.490.030.020.610.56
CEM0.620.650.641.17−0.12
MF0.560.640.641.23−0.05
TCIMF0.480.590.581.09−0.08
0.6 mTUTDF0.890.690.111.690.89
OSP0.890.380.381.370.61
SAM0.740.350.241.180.59
ACE0.510.070.050.640.52
CEM0.600.520.501.060.04
MF0.530.510.501.020.01
TCIMF0.280.530.511.050.01
0.9 mTUTDF0.760.540.171.540.83
OSP0.740.600.481.380.30
SAM0.640.280.250.900.37
ACE0.510.070.060.610.48
CEM0.640.560.511.160.09
MF0.560.550.511.140.08
TCIMF0.280.480.480.990.03
The bold entries represent the best performance in each column at a certain depth.
Table 7. Execution time of different methods.
Table 7. Execution time of different methods.
DatasetExecution Time (Second)
TUTDFMFCEMACEOSPSAMTCIMF
Iron0.6881.0681.1321.3750.9121.4701.110
Stone0.6891.0921.1101.3720.9301.2721.076
Rubber0.6871.0721.2101.3740.8901.4281.100
The bold entries represent the best performance in each row.
Table 8. Execution time of different modules.
Table 8. Execution time of different modules.
DatasetExecution Time (Second)
TUTDFSpatialDomainDetection
Iron0.6880.0390.3480.301
Stone0.6890.0400.3450.304
Rubber0.6870.0390.3510.297
Table 9. Comparison of effect with and without domain adaptive, ‘Yes’ is with domain adaptive, ‘No’ is without domain adaptive.
Table 9. Comparison of effect with and without domain adaptive, ‘Yes’ is with domain adaptive, ‘No’ is without domain adaptive.
Test Data AUC ( D , F )
YesNo
Iron (1.6 m)1.000.88
Stone (0.5 m)0.980.78
The bold entries represent the best performance in each row.
Table 10. Interval settings of domains for different materials.
Table 10. Interval settings of domains for different materials.
Iron D 1 [ 0 , 0.4 ) D 2 [ 0.2 , 0.6 ) D 3 [ 0.4 , 0.8 ) D 4 [ 0.6 , 1.4 ) D 5 [ 1.2 , 2.0 ) D 6 [ 1.8 , 2.5 ]
Stone D 1 [ 0 , 0.4 ) D 2 [ 0.2 , 0.6 ) D 3 [ 0.4 , 0.8 ) D 4 [ 0.6 , 1.2 ) D 5 [ 0.9 , 1.7 ) D 6 [ 1.4 , 2.0 ]
Rubber D 1 [ 0 , 0.3 ) D 2 [ 0.2 , 0.5 ) D 3 [ 0.4 , 0.7 ) D 4 [ 0.6 , 0.9 ) D 5 [ 0.8 , 1.1 )
The unit of measurement is meter (m).
Table 11. Real and estimated depths of underwater targets of different materials. ‘R’ denotes ‘Real’, ‘E’ denotes ‘Estimated’.
Table 11. Real and estimated depths of underwater targets of different materials. ‘R’ denotes ‘Real’, ‘E’ denotes ‘Estimated’.
Iron(R)0.1000.5001.0001.600
Iron(E)0.0510.4450.9781.551
Stone(R)0.4000.7001.6001.900
Stone(E)0.3500.6231.5771.875
Rubber(R)0.1000.3000.6000.900
Rubber(E)0.1210.3170.6100.903
The unit of measurement here is meter (m).
Table 12. The effect of the spatial–spectral process on the detection of iron plate: ‘Yes’ is with spatial–spectral process, ‘No’ is without spatial–spectral process.
Table 12. The effect of the spatial–spectral process on the detection of iron plate: ‘Yes’ is with spatial–spectral process, ‘No’ is without spatial–spectral process.
Depth0.1 m0.5 m1.0 m1.6 mAverage
NoRemotesensing 15 01023 i001Remotesensing 15 01023 i002Remotesensing 15 01023 i003Remotesensing 15 01023 i004
A U C ( D , F ) 1.001.001.000.990.9975
YesRemotesensing 15 01023 i005Remotesensing 15 01023 i006Remotesensing 15 01023 i007Remotesensing 15 01023 i008
A U C ( D , F ) 1.001.001.001.001.00
The bold entries represent the best performance in each row.
Table 13. The effect of the spatial–spectral process on the detection of stone plate.
Table 13. The effect of the spatial–spectral process on the detection of stone plate.
Depth0.4 m0.7 m1.6 m1.9 mAverage
NoRemotesensing 15 01023 i009Remotesensing 15 01023 i010Remotesensing 15 01023 i011Remotesensing 15 01023 i012
A U C ( D , F ) 1.001.000.780.770.8875
YesRemotesensing 15 01023 i013Remotesensing 15 01023 i014Remotesensing 15 01023 i015Remotesensing 15 01023 i016
A U C ( D , F ) 1.001.000.870.780.9125
The bold entries represent the best performance in each column.
Table 14. The effect of the spatial–spectral process on the detection of rubber plate.
Table 14. The effect of the spatial–spectral process on the detection of rubber plate.
Depth0.1 m0.3 m0.6 m0.9 mAverage
NoRemotesensing 15 01023 i017Remotesensing 15 01023 i018Remotesensing 15 01023 i019Remotesensing 15 01023 i020
A U C ( D , F ) 0.980.970.840.710.87
YesRemotesensing 15 01023 i021Remotesensing 15 01023 i022Remotesensing 15 01023 i023Remotesensing 15 01023 i024
A U C ( D , F ) 0.990.980.880.760.90
The bold entries represent the best performance in each column.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Li, J.; Zhang, P.; Zheng, L.; Shen, Y.; Li, Q.; Li, X.; Li, T. A Transfer-Based Framework for Underwater Target Detection from Hyperspectral Imagery. Remote Sens. 2023, 15, 1023. https://doi.org/10.3390/rs15041023

AMA Style

Li Z, Li J, Zhang P, Zheng L, Shen Y, Li Q, Li X, Li T. A Transfer-Based Framework for Underwater Target Detection from Hyperspectral Imagery. Remote Sensing. 2023; 15(4):1023. https://doi.org/10.3390/rs15041023

Chicago/Turabian Style

Li, Zheyong, Jinghua Li, Pei Zhang, Lihui Zheng, Yilong Shen, Qi Li, Xin Li, and Tong Li. 2023. "A Transfer-Based Framework for Underwater Target Detection from Hyperspectral Imagery" Remote Sensing 15, no. 4: 1023. https://doi.org/10.3390/rs15041023

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop