Next Article in Journal
Digital Quantum Simulation of the Spin-Boson Model under Markovian Open-System Dynamics
Next Article in Special Issue
BPDGAN: A GAN-Based Unsupervised Back Project Dense Network for Multi-Modal Medical Image Fusion
Previous Article in Journal
The Diagnosis of Congestive Heart Failure Based on Generalized Multiscale Entropy-Wavelet Leaders
Previous Article in Special Issue
Infrared and Visible Image Fusion for Highlighting Salient Targets in the Night Scene
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Water-Air Interface Imaging: Recovering the Images Distorted by Surface Waves via an Efficient Registration Algorithm

1
School of Information and Communication, Guilin University of Electronic Technology, Guilin 541000, China
2
School of Artificial Intelligence, Hezhou University, Hezhou 542800, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(12), 1765; https://doi.org/10.3390/e24121765
Submission received: 9 October 2022 / Revised: 28 November 2022 / Accepted: 29 November 2022 / Published: 2 December 2022
(This article belongs to the Special Issue Advances in Image Fusion)

Abstract

:
Imaging through the wavy water–air interface is challenging since the random fluctuations of water will cause complex geometric distortion and motion blur in the images, seriously affecting the effective identification of the monitored object. Considering the problems of image recovery accuracy and computational efficiency, an efficient reconstruction scheme that combines lucky-patch search and image registration technologies was proposed in this paper. Firstly, a high-quality reference frame is rebuilt using a lucky-patch search strategy. Then an iterative registration algorithm is employed to remove severe geometric distortions by registering warped frames to the reference frame. During the registration process, we integrate JADE and LBFGS algorithms as an optimization strategy to expedite the control parameter optimization process. Finally, the registered frames are refined using PCA and the lucky-patch search algorithm to remove residual distortions and random noise. Experimental results demonstrate that the proposed method significantly outperforms the state-of-the-art methods in terms of sharpness and contrast.

1. Introduction

Imaging through water is an important research direction of underwater optics and marine optics, which is widely employed in environmental, economic and military application fields [1,2,3,4,5,6]. Unlike other underwater imaging systems, the main challenge in this imaging scenario comes from the air–water interface. Random fluctuations of surface waves will introduce a variety of effects, such as non-single viewpoint distortion [7], double images [8], scene distortion, illumination caustics [9] and black holes [10,11]. These effects can cause extreme distortion and motion blur in the images. Therefore, inverting these distortions and retrieving high-quality images remains a formidable task.
Several efforts have been made in earlier works to address the issue of retrieving the distortion-free image from a video sequence distorted by surface waves. If the turbulence of surface waves is small, averaging the sequence or the blind deconvolution method [12] will perform well, whereas the reconstruction result using them will be extremely blurry once the waves are strong. In prior research, better techniques were to select the image patches with the least distortions and combine them [13,14,15,16,17,18]. The limit of these methods lies in the fact that they assume the corresponding patches from different times are spatially invariant, which causes such algorithms to handle only slightly fluctuating situations and is easy to cause information loss. Later, another distortion mitigation approach, an image restoration scheme of water surface modeling based on image sequence is described in [19,20,21,22,23,24]. However, it is challenging to portray different complicated fluid motion events in real environments using such a simplistic spatial model.
Recently, registration technology, which was initially developed to recover atmospheric turbulence images, has been applied to this problem. The typical strategy [25,26,27] is to use the temporal mean as the reference frame and to improve mean–frame correlation by blurring the sharp frames with a blur kernel. The mean and frames are then iteratively refined via robust registration. Although this scheme can significantly eliminate severe geometric distortions, it is prone to inadvertent misalignment and motion blur. Once in a strong wave environment, such effects will be more pronounced. Subsequently, Zhen Zhang et al. [28,29] attempt to reconstruct a higher-quality reference frame by using a lucky-patch search and a deblurring process. they attempted to guide the registration of a sharper boundary to remove geometry distortions. However, they ignore two key factors that affect the registration process: (1) The premise of a lucky patches search is that the corresponding regions from different frames are spatially invariant or approximate; otherwise, it will result in the apparition of double images in the synthetic frame; (2) The deconvolution procedure required in deblurring is often ill-posed, even the most advanced deblurring algorithms, such as [30,31,32], frequently fail and may result in the appearance of visual artifacts (commonly described as “artificial” edges) [33]. Although the contrast of the image is improved in their processing results, the overall performance evaluation is not satisfactory, and even inferior to Oreifej’s method [25] in individual indicators.
In this work, we present a video-sequence-based method for image reconstruction. At first, a new lucky-patch search method is used to reconstruct a high-quality reference frame, where all warped frames are aligned in the first iteration, followed by the patch search in subsequent iterations. Secondly, the warped frames are then registered against the reference frame using image registration to eliminate the severe geometry distortions in the images. Finally, the registered frames are refined using PCA and the lucky-patch search algorithm to remove residual distortions and random noise.It is worth noting that in the registration stage, we introduce a hybrid optimization strategy to speed up the optimization process of the control parameters. The experimental results show that the proposed approach outperforms the state-of-the-art methods in terms of sharpness and contrast.

2. Related Work

Researchers from various fields have attempted to tackle the problem of recovering the real scene from a distorted video sequence. According to Cox–Munk law [34], the normals of the water surface generally follow a Gaussian distribution when the interface is sufficiently large and calm. As such, numerous approaches related to the lucky-patch were presented Efors et al. [13] described the reconstruction issue as a manifold embedding problem in which a patch with the smallest global distance was chosen as the distortion-free patch. Donate et al. [14] presented a multi-stage clustering approach to eliminate regions with significant amounts of translation and motion blur, independently. Wen et al. [15] used a bi-spectral analysis approach to transform the reconstruction issue into a phase recovery issue. A.V. Kanaev et al. [16] developed a method for image restoration based on optical flow and the lucky region. The lucky patch was determined using an image evaluation metric and the current frame’s nonlinear gain coefficient at each location. Later in [17], A.V. Kanaev et al. propose a structural tensor-oriented image quality measure (STOIQ) to evaluate image quality. Zhang et al. [18] recently proposed “center evolution imaging”, a simple yet effective progressive recovery method that improves restoration quality by constantly updating following warped frames. Subsequently, several researchers attempted to recover the distortion-free image based on model methods. Tian et al. [19] used a model-based tracking approach to reconstruct the real scene and instantaneous shape of the water surface simultaneously. In later work, they proposed a data-driven iterative algorithm [20] and introduced the operation of “pulling back”. In Ref [21], Seemakurthy K et al. proposed methods to remove motion blur caused by unidirectional cyclic waves and circular ripples, respectively. More recently, Li et al. [22] demonstrated their trained convolutional neural network that can remove dynamic refraction using a single image. James et al. [26] came up with a method for restoring underwater images using both compressive sensing (CS) and a local polynomial image representation. Thapa et al. [24] addressed a distortion-guided network (DG-Net) to restore distortion-free underwater images. They first use a physically constrained convolutional network to estimate the distortion map from the refracted image. Then, using the distortion map as a guide, a generative adversarial network was used to retrieve a sharp image with no distortion.
Recently, another typical approach, registration technology, was used to tackle this issue. Oreifej et al. [25] proposed a two-stage restoration method that first removed structured noise using a robust registration process, and then eliminated sparse noise by rank minimization. Hu Wenrui et al. [26] proposed a motion field kernel regression-based method to reconstruct distortion-free images. Halder et al. [27] introduced a pixel-shift map-based restoration method. First, the method uses image registration to estimate pixel displacement maps of warped images, and then reconstructs distortion-free scenes based on these maps. Zhen Zhang et al. [28,29] attempted to reconstruct a reference frame using a lucky-patch search and a deblurring process. They tried to guide the distorted frame registration to sharper boundaries to remove geometric distortions, but the results were not convincing. Cai et al. [35] proposed a reconstruction algorithm for recovering distorted images based on a two-round registration algorithm, which obviously improves the restoration accuracy but reduces the computational efficiency of image registration.
With the aim to raise the restoration accuracy, visual effects, and computational efficiency, we propose a new image reconstruction approach that utilizes a high-quality reference image to guide the distorted frames toward focusing on the sharper boundary or region. Moreover, an optimization strategy which combines the JADE with LBFGS methods is introduced to accelerate the iterative registration process. Unlike other methods, ours does not need a particular camera height [19], image priori [20], or special illumination [36,37]. Our approach is similar to the state-of-the-art [25,28,29,30] in that it only needs a short sequence (61 frames) rather than 800 frames in [14], 300k frames in [22], 100 frames in [23], and 43,600 frames in [24], where it is the prerequisite for underwater imaging systems to achieve covert observation.

3. Methods

Assume a water-deformed video sequence V = { I 1 , I 2 , , I n } in which each frame I k   R w × h ,   k = 1 , , n is warped by an unknown deformation Γ ( x ) . Our primary objective is to rebuild a distortion-free frame sequence V f = { I f 1 , , I f n } and a high-quality dewarped frame I F . The framework of the image reconstruction method in this paper is shown in Figure 1.

3.1. Principles

As Figure 1 shows, we optimize both the reference frame and registered frames using a robust iteration registration procedure, in which at each iteration the distorted frames are registered against the reference frame which is closer to the real-scene image. At first, a new lucky-patch search strategy is proposed to generate a higher-quality reference frame than the mean, in which we align all distorted frames to the average frame using an initial registration in the first iteration, the purpose of which is to remove the spatial variability between images introduced by random fluctuations of the water surface. We then perform the patch search algorithm in subsequent iterations. Next, similar to [25], we estimate the blur kernel to make sure that the distorted frames have the same blur level as the reference frame. Subsequently, a non-rigid image registration based on the hybrid optimization strategy is employed to estimate the motion vector field Γ ( x ) and remove the serious geometric distortions. The hybrid optimization strategy consists of a global optimization algorithm JADE [38] and a local optimization algorithm LBFGS [39], with the purpose of avoiding deformation parameters falling into local optimum and improve the registration process.
After a few iterations, the robust iteration registration will converge to new registered frames V r = { I r 1 , I r 2 , , I r n } which are less distorted and blurred. However, there are still several sparse random noise and misalignments among them, which are difficult to remove by robust registration. In order to acquire a distortion-free sequence V f = { I f 1 , , I f n } and a high-quality dewarped frame I F , the registered frames are refined using PCA technology and the lucky-patch search method described above to remove residual distortions and random noise.

3.2. Lucky-Patch Search with Initial Registration

In order to reconstruct a higher-quality synthetic image than the temporal mean of the sequence, we introduce an image-quality-assessment-driven lucky-patch search approach. Firstly, the video frames are divided into N overlapping sub-regions with the same size, R k = P 1 , P 2 , , P N ,   k = 1 , , n   , where the overlapping area of adjacent image patches is 50 % . The k-Medoids clustering algorithm [41] is then used to try to search for the set of high-quality patches with the fewest distortions from each patch sequence P m = b m 1 , , b m n   , m = 1 , , N . Subsequently, all of the high-quality patches from each patch sequence are averaged into a single patch. Finally, a complete fused image is constructed by splicing all the average patches together.
In this paper, we use edge strength similarity (ESSIM) [42] as a clustering criterion. ESSIM is a structural similarity index which has better performance than other image quality metrics such as SSIM [43] and FSIM [44]. EESIM is used for measuring the similarity between the patch b m k   ,   m = 1 , , N   ;   k = 1 , , n and the mean u m of the corresponding patch sequence. EESIM is given by
ESSIM u m , b m k = 1 M η = 1 M 2 E u m , η E b m k , η + c E u m , η 2 + E b m k , η 2 + c
where M denotes the total number of pixels. E u m , η and E b m k , η represent the total edge-strength and the edge-strength of b m k at the η   t h pixel, respectively. The c is the scaling parameter.
Furthermore, in order to ensure the corresponding patches from different times are spatially invariant or approximate, we first register all frames to the mean at the 1 t h iteration. The framework to reconstruct the reference frame is shown in Figure 2.

3.3. Image Registration Based on the JADE & LBFGS Optimization Strategy

Similar to [30], a nonrigid image registration algorithm based on the JADE and LBFGS optimization strategies is applied to estimate the deformation Γ ( x ) and remove the distortion. Assume the image domain as Ω = ( x , y )   |   0 x   <   X ,   0 y < Y , the goal of the iterative registration is to find the optimal transformation Γ ( x ) to guide registering the video sequence to the reference frame. The deformation Γ ( x ) can be expressed as
Γ ( x ) = Γ local ( Γ global ( x ) )
where Γ global and Γ local are the global transformation and the local transformation, respectively. Γ global describes the overall motion of the moving image, which is generally represented by an affine transformation with six parameters, namely
Γ global ( x   ;   Θ ) = θ 11 θ 12 θ 21 θ 22 x y + θ 13 θ 23
where Θ = θ 11 , θ 12 , θ 13 , θ 21 , θ 22 , θ 23 represents the set of the six parameters. Γ local models the local deformation of the distorted frames. Local deformation is difficult to characterize using parameterized transformations since the wavy water surface introduces anisoplanatism effects for imaging through water. Instead, we employ the B-spline based free-form deformation model (FFD) to represent the local deformation. The basic ideal of FFD is to describe the geometry distortions between the moving image and the reference frame by manipulating an underlying mesh of control points. Let Φ as a n x × n y mesh of control points φ i , j with uniform spacing, the FFD can be expressed as
Γ local x   ;   Φ = t = 0 3 l = 0 3 B t ( u ) B l ( v ) φ i + t , j + l
where i = x / s x 1 ,   j = y / s y 1 ,   u = x / s x x / s x ,   v = y / s y y / s y and B l is a standard B-spline basic function which defines the weight of the control points contributing in the deformation of the arbitrary point x in the image. The B-spline basic functions are given by
B 0 ( v ) = ( 1 v ) 3 / 6 B 1 ( v ) = ( 3 v 3 6 v 2 + 4 ) / 6 B 2 ( v ) = ( 3 v 3 + 3 v 2 + 3 v + 1 ) / 6 B 3 ( v ) = v 3 / 6
We minimize a cost function associated with the global transformation parameters Θ and the local control parameters Φ to determine an optimal transformation Γ ( x ) . The function can be defined as
E Θ , Φ = C similarity Θ , Φ + λ C smooth Φ
where C similarity Θ , Φ is the similarity metric function which describes the difference between the reference frame I R and the distorted frame I k . Here, we utilize the squared sum of intensity differences (SSD) to denote the similarity metric function, namely
C similarity Θ , Φ = x Ω I k Γ x I R x w × h
where I k ( Γ ( x ) ) and I R ( x ) are the intensity of the corresponding pixels of two images, respectively. Furthermore, λ is the weighting parameter which defines the tradeoff between the alignment of the two images and the smoothness of the transformation. C smooth ( Φ ) is a regularization penalty term that controls the smoothness of the transformation. The penalty term in 2-D is expressed as follows:
C smooth Φ = 1 A 0 X 0 Y 2 Φ x 2 2 + 2 Φ y 2 2 + 2 2 Φ x y 2   d x d y
where A denotes the image domain.
Unlike the previous works [25,26,27,28,29,30], we no longer use the LBFGS search algorithm or the gradient descent method alone to minimize Equation (6), but combine the JADE and LBFGS algorithms as the search strategy that is described in detail in Section 3.3.1.

3.3.1. Optimization Strategy

Consider the minimization of a smooth nonlinear function f :   τ ,
min f δ
where δ denotes the multi-variable vector, δ τ , and τ is the length of variables.
First of all, first-order techniques, such as gradient descent, can be utilized to solve Equation (9). However, gradient descent may not perform well in ill-conditioned situations [45], and the convergence speed of the method will be slower as it approaches the objective function’s lowest point. Typically, second-order techniques, such as Newton’s method, are used to address this deficiency. Unfortunately, Newton’s method needs the inverse of the Hessian matrix 2 f ( δ ) at each iteration, and its computing cost is prohibitively expensive for variables with a large dimension. In quasi-Newton approaches such as [39,46,47], an approximated Hessian matrix is used to take advantage of the curvature while also reducing computing complexity. Generally speaking, BFGS performs the best in practice. The limited-memory variant of the original BFGS algorithm (LBFGS) reduces the computational cost of each iteration even more, making it one of the most effective algorithms in this domain. However, similar to other local optimization methods, the LBFGS algorithm easily converges to the local optimal solution once the initial value δ 0 is not selected properly. Therefore, how to select the initialization parameters becomes the key to obtaining the optimal parameters.
JADE [38] is a simple and effective global optimization method, which is a variant of the differential evolution algorithm. Differential evolution follows the general process of evolutionary algorithms. The initial population { δ z 0 = ( δ z , 1 0 , δ z , 2 0 , δ z , τ 0 ) |   z = 1 , 2 , , N P } is randomly generated according to a uniform distribution, where N P is the population size. If nothing is known about the solution of the objective function, the initial population is randomly selected. If a preliminary solution is available, the initial population is usually generated by adding random deviations from the normal distribution. The JADE algorithm then enters a cycle of evolutionary operations: mutation, crossover, and selection. After repeating the above process, the algorithm will gradually approach the global optimal solution.
Therefore, in order to avoid the transformation parameters ( Θ , Φ ) falling into a local optimum, we attempt to minimize Equation (6) by utilizing JADE with LBFGS algorithms. Among them, JADE is used to calculate a neighboring solution to the global optimal solution, and LBFGS is used to obtain the optimal solution. Furthermore, for computational efficiency, we do not use an iterative multi-resolution search strategy [40,41] any longer in the stage where the affine transformation parameters Θ are optimized. In practice, it is sufficient to utilize the combination optimization strategy.
Specifically, the optimization process in this paper can be divided into the following two stages. In the first stage, the integrated JADE and LBFGS algorithms are used to optimize the global transformation parameters Θ . In the second stage, the local transformation parameters Φ are optimized through a hierarchical multi-resolution strategy [40] such that the resolution of the control mesh is increased, along with the image resolution in a coarse-to-fine fashion.

3.4. Post-Processing

After a few iterations, we can obtain a new image sequence V r = { I r 1 , I r 2 , , I r n } which is less distorted and blurred. However, there are still several sparse random noise and misalignments among them, which are difficult to remove by robust registration. In order to obtain a distortion-free image sequence V f = { I f 1 , , I f n } , we refine the image sequence using PCA technology [48]. Subsequently, we perform the patch search strategy described in Section 3.2 again to select the patches with less distortion from the registered frames to fuse a higher-quality single image I F . Unlike the first fusion, we first divide each patch sequence into five groups rather than two groups using the k-Medoids clustering algorithm. The reason for this is that the differences between adjacent images are already small after iterative robust registration and sparse denoising. Moreover, to improve the registration efficiency, we can add a suitable termination condition ξ ( v ) to determine if the registration process has ended or not. The comprehensive process for image reconstruction is described in Algorithm 1.
Algorithm 1: The correct format is:An efficient method for recovering the images distorted by surface waves
Input: Distorted Video sequence
   V = { I 1 , I 2 , , I n }   ,   I k w × h , k = 1 , 2 , , n
Output: Distortion-free sequence and high-quality frame
   V f = { I f 1 , , I f n }   ,   I f k w × h , k = 1 , 2 , , n   ;   I F w × h
While   r m a x i t e r a t i o n   L do
if r = 1
I R T e m p o r a l M e a n ( V ) ;
else
I R P a t c h e s _ s e a r c h ( V ) ;
end
for k = 1 : n
   B E e s t i m a t e B l ur K e r n l e ( Γ ) ;
I b l u r G a u s s i a n F i l t e r ( I k , B ) ;
Γ C o m p u t e r D e f o r m a t i o n ( I b l u r , I R ) ;
V r ( : , : , k ) D w r a p I k , Γ ;
end
V = { I 1 , I 2 , , I n } V r = { I r 1 , I r 2 , , I r n } , I r k w × h , k = 1 , 2 , , n ;
r = r + 1 ;
end
V f = { I f 1 , , I f n }   P C A V r   , r = L
Computer the frame I F
for each patch sequence P m = b m 1 , , b m n   , m = 1 , , N
Q = q 1 , , q n C o m p u t e E S S I M ( P m )
P b e s t = { b m 1 , b m 2 , , b m c } K m e d o i d C l u s t e r ( P m , Q ) ,   1 c n
u m T e m p o r a l M e a n ( P b e s t )
end
U = u 1 , u 2 , u m , m = 1 , 2 , , N
I F   F u s i o n ( U )

4. Results

In the experiment, the proposed method described above was implemented on MATLAB (Math Work Co., Natick, MA, USA). To validate the effectiveness of our method, we first conducted extensive experiments using several standard underwater sequences from [19]. Then, we further tested the proposed method with two typical real-scene sequences from [18]. Moreover, we made a comparison between our experimental results with other state-of-the-art methods, such as Tian’s method [19], Oreifej’s method [25], Zhen Zhang’s method [28] and Tao Sun’s method [29]. The maximum number of iterations for all image registration algorithms was set as five and each patch of the corresponding sequence was set in the same size. The source codes of our method are available online in [49].

4.1. Test with the Data Set from the Air-to-Water Imaging Scenario

To more accurately validate the proposed method, the experiment employed the same standard underwater scene sequences from [19]. The data sets contain five image sequences (as Table 1 shows) of which each sequence contains 61 frames. Table 1 shows the size of each frame and patch utilized in the patch search procedure.
In order to quantify the results of these methods, we employed three standard image quality/similarity metrics for quantitative evaluation, namely PSNR [30], MSE [6] and SSIM. Above metrics are classified into full-reference that represent the different between the moving frame and the reference frame. The expressions for MSE, PSNR, and SSIM can be described as
MSE = 1 w × h x = 1 w y = 1 h F ( x , y ) G ( x , y ) 2
PSNR = 10 log 10 ( max ( G ) 2 1 w × h x = 1 w y = 1 h F ( x , y ) G ( x , y ) 2
SSIM = ( 2 u F u G + c 1 ) ( 2 σ F G + c 2 ) ( u F 2 + u G 2 + c 1 ) ( σ F 2 + σ G 2 + c 2 )
where F and G are the estimated image and the reference image, respectively. max ( · ) represents the maximum value. u F and u G are the averages of F and G , respectively. σ F and σ G denote the variance of F and G , respectively. σ F G is the covariance of F and G . c 1 = (   0.01 × β   ) , c 2 = (   0.03 × β   ) denote the constants, in which β represents the dynamic range of pixel values. Moreover, since the template image for “Checkerboard” and “Large fonts” are unknown, a novel non-reference underwater image quality measurement method UIQM [50] is used as another quality metric. The UIQM metric is given by
UIQM = λ 1 × UICM + λ 2 × UISM + λ 3 × UIConM
where UICM, UISM and UIConM represent the image colorfulness metric, the image sharpness metric and the image contrast metric, respectively. λ 1 , λ 2 , λ 3 are constants which are generally set as 0.0282, 0.2953 and 3.5753. Due to the image sequences being gray images, the UICM is identically equal to zero.

4.1.1. Analysis of Restoration Results

Figure 3 shows the recovery results of our method and other state-of-the-art methods. The image quality evaluation results are shown in Table 2. Compared with Oreifej’s results, our results perform better in term of sharpness and contrast and preserve more detail of the underwater scene, which illustrate that the higher-quality reference frame than the mean better guides the warped image to focus on sharp boundaries or areas. In addition, we find that the results from Z. Zhang and T. Sun’s methods still contain lots of residual distortions such as misalignment, ghosting and local deformation, etc. The main reason is that they ignore the premise of spatial invariance, which is very important for the lucky-patch search strategies. Unlike the methods [28,29], we attempted to reconstruct a high-quality reference frame utilizing a new patches search strategy which added an alignment process at the first iteration. The results show that our method can reconstruct all of the underwater scenes and make high-quality results that are better than others.

4.1.2. Analysis of Computational Efficiency

In term of computational efficiency, we compared the running time of different algorithms using two different termination conditions. In order to ensure fairness, we performed these methods on the same laptop computer. The operating system contains a CPU (i7-6700HQ) and RAM (8 GB). The data set was “Middle fonts”. Moreover, in Oreifej’s work [25], it was proved that the processing time of Tian’s method was twice that of Oreifej’s method, so it is no longer compared with Tian’s method here in terms of computational efficiency.
Firstly, the termination condition is set as the maximum number of iterations. During the registration process, the comparison of the running times of different algorithms is shown in Figure 4. The average iterative times of different methods are 261.6 s, 236.4 s, 245 s, and 187.6 s, respectively. Because the length of the data set is 61, the average cost times of each frame are 4.29 s, 3.88 s, 4.02 s and 3.07 s, respectively. The results demonstrate that the proposed method may expedite image registration and shorten image processing time.
For the purpose of raising efficiency, we also consider another approach, which is to set the iteration termination condition as the difference between the registered frames and the average of the current sequence V . Similar to [25], the termination condition can be expressed as
Τ V = k x I k ( x ) I m e a n ( x ) w × h × n
where the threshold is set to 0.025.
The restoration time of different methods is shown in Figure 5. We can notice that our method’s overall running time and iteration numbers are significantly less than those of other approaches. For Oreifej’s results, the main reason is due to the severely blurred mean image. The blurring of the mean will become worse once the fluctuation of the water surface becomes stronger. In order to remove geometry distortions, the video frames are then also blurred to the same fuzzy level. However, the above strategy may reduce the registration efficiency due to the large amount of overlap and ghosting present in the mean image. As for Zhen Zhang’s and Tao Sun’s methods, the main reason is due to two aspects: (1) they ignored the effect of spatial inconsistency between different frames on the fused image. (2) They did not consider the problem of artificial artifacts introduced by deblurring methods. These problems can easily lead to distorted images being registered to the wrong sharp edges, which often have priority in their strategy. Thereby, a new patches search strategy is applied to generate the reference frame in our method, where we first perform an alignment operation for the input video sequence. Our reference frame is better than the mean image in terms of sharpness and contrast. Moreover, another important reason is that our combinatorial optimization strategy is significantly better than other algorithms in term of computational efficiency, which we discuss in detail in Section 5.3.
Therefore, the method has better computational efficiency when the number of iterations is the same or unknown, which helps to promote the application of this technique in underwater covert observations, such as a virtual periscope.

4.2. Test with the Data Set from the Water-to-Air Imaging Scenario

In order to further confirm the effectiveness of the proposed approach,, we tested data sets from another imaging scenario, the underwater-to-air imaging scenario. The data sets come from Zhang’s experiment [18], whose source data are also available online. The data sets contain two typical sequences, “gravity” and “ripple”, in which the first sequence is distorted by a gravity wave and the other is distorted by wind-generated capillary waves. Each sequence is also composed of 61 frames. The maximum number of iterations for all image registration algorithms was set to five, and each patch in the related sequence was set to the same size.
The recovery results with different methods are shown in Figure 6. Moreover, we compared these algorithms with a non-reference quality evaluation metric, UIQM, considering that template images are not available for the above datasets. The comparison results of the UIQM metric for different methods are shown in Table 3. The results show that our algorithm is still effective in water-to-air imaging scenarios and produces higher-quality results.

5. Discussion

5.1. Analysis of the Lucky-Path Search Strategy

5.1.1. Selection the Image Quality Metric

During the patch fusion process, ESSIM was chosen as the quality metric. We also tested our algorithm using SSIM, FSIM, and NIQE [51] metrics, where NIQE is a non-reference metric. The fusion results based on the different image quality metrics at the 2th iteration stage are shown in Figure 7. In addition, the quality scores for the above four quality indicators (here, SSIM is used as the evaluation standard) are 0.6500, 0.6571, 0.6510, and 0.6289, and the computational efficiency is 4.56 s, 11.03 s, 78.12 s, and 33.03 s, respectively. According to the above results, considering the computational efficiency of the algorithm, we chose EESIM as the image quality metric at the patch fusion stage. The results show that other reliability metrics can be considered in future work to further optimize our patch fusion process.

5.1.2. Analysis of the Evolution of Reference Frame

Figure 8 demonstrates the evolution of the reference frame for the different methods in the registration process. The image quality evaluation results for the reference frames of different algorithms for each iteration are shown in Table 4 In Z. Zhang and T. Sun’s approaches [28,29], they directly perform the patch search process to generate reference frames, ignoring the issue of spatial differences in input sequences. For this reason, the reference frames inevitably produce ghosting, misalignment, and overlapping effects as shown in Figure 8 (remarked by the red rectangle). Furthermore, we find that there are artifacts in the reference frame that are not related to image details in the results of Z. Zhang’s method, which are clearly the result of blind convolution. In other words, deblurring operations similar to blind convolution algorithms are obviously infeasible and tend to degrade the restoration results. However, our method can always generate higher-quality reference frames than the mean, which would guide the registration process and lead to better restoration results, as shown in Figure 3. The reference frames generated by our patch search strategy perform better in sharpness and contrast than other methods. As shown in Figure 8, compared with other methods, the text information in the reference frame reconstructed by our method is clearer and more accurate (remarked by the green rectangle). The results show that it is essential to align the input sequence before applying the lucky block fusion technique. Obviously, our patch search strategy always produces higher-quality results.

5.2. Analysis on the Choice of Reference Frame Deblurring or Frame Blurring

Figure 9 demonstrates the evolution process of the synthetic frame I F for the “Middle fonts” sequence in the iterative registration process with three cases. As shown in Figure 9, our method recovers high-quality results after several iterations under all three strategies. However, textual details in images, such as “Imaging”, “Tracking”, and “computer”, can only be clearly reconstructed using our method.
Furthermore, it is worth noting that, compared to the first strategy, both frame blurring and reference frame deblurring strategies can produce sharper results, which seems to mean that we can try to seek a more effective deblurring scheme to obtain high-quality results.

5.3. Comparison of the Optimization Strategy

The comparison results of the running time of different optimization strategies after each iteration are shown in Figure 10. The results show that our optimization strategy outperforms the state-of-the-art optimization methods used in [25,26,27,28,29,30,41] in terms of computational efficiency. The main reasons can be attributed to the following two aspects: (1) The combined strategy of JADE and LBFGS is used to estimate the initial value of the optimization parameters ( Θ , Φ ) , which avoids the local optimum problem. (2) In the affine parameter optimization stage, the multi-resolution optimization strategy is no longer used.

5.4. Compared with Deep Learning Method

To more thoroughly validate its performance, we compare our method to another cutting-edge strategy, deep learning techniques, which Tian’s method [19] also belongs to. In terms of repeatability, James’ technique [23], the source code for which is available online, was chosen as a reference method. The proposed method and James’ method were tested using identical testing data, “large fonts” and “ripple.” Figure 11 shows the restoration effects of the two methods. The results show that these two methods can effectively reduce the geometric distortion and motion blur of the image and restore better results. However, in terms of contrast and sharpness, our algorithm shows better performance. As Figure 11 shows, when the water surface disturbance is small, both James’ method and our method can restore the distorted scene image well, and our algorithm shows better sharpness (as indicated by the green markers). Once the water surface disturbance becomes stronger, the James method will not be able to recover high-quality results, especially the details of the scene (marked by the red rectangle). However, our method can reconstruct better boundaries and more complete details. Experimental results show that simple spatial distortion models similar to [23] may fail in heavily disturbed environments. However, our algorithm still produces high-quality results.

6. Conclusions

This paper proposes a new image reconstruction scheme with distorted video sequences while keeping image restoration accuracy and computational efficiency in mind. The main contributions are as follows: (1) The optimal combination of lucky- patch fusion and iterative registration algorithms is realized, which effectively eliminates the ghosting problem in a fused reference frame and raises the recovery accuracy of the distorted frames; (2) Introduce a mixed optimization strategy of JADE and LBFGS to accelerate the registration process.
We extensively experimented on the proposed method using five standard underwater sequences from [18] and the two typical data sets from [19]. Experiments show that our approach performs better in terms of sharpness and contrast. Therefore, this means that our algorithm has better development prospects in applications such as underwater or airborne objective monitoring, obstacle avoidance for underwater vehicles, seafloor mapping, etc.
To further raise the recovery accuracy and computational efficiency, we will develop a more effective method to generate a higher-quality reference frame in the future. Moreover, it is worth noting that the machine learning methods [53,54,55,56,57,58] have unique advantages in improving image recognition. For application scenarios, where covert monitoring is not considered, once the model is established, it can be used as a means of quickly detecting objects.

Author Contributions

All authors have made significant contributions to this paper. B.J. designed the research, conducted experiment analysis and wrote the manuscript; C.M. provided guidance; D.Z. conducted data curation; Q.H. conducted validation; J.A. provided supervision and funding. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Guangxi National Science Foundation, grant number 2018GXNSFAA294056; the Guangxi Young and Middle-Aged Teachers’ Basic Research Ability Improvement Project, grant number 2022KY0703; the School-level Scientific Research Project of Hezhou University, grant number 209200297; and the Achievements of Hezhou University’s 14th Five-Year “need” Level Discipline “Information and Communication Engineering” Construction Project.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

All data or codes used to support the findings of this study are available from Ref [49].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Molkov, A.A.; Dolin, L.S. The Snell’s window image for remote sensing of the upper sea layer: Results of practical application. J. Mar. Sci. Eng. 2019, 7, 70. [Google Scholar] [CrossRef] [Green Version]
  2. Gardashov, R.H.; Gardashov, E.R.; Gardashova, T.H. Recovering the instantaneous images of underwater objects distorted by surface waves. J. Mod. Opt. 2021, 68, 19–28. [Google Scholar] [CrossRef]
  3. Li, N.; Thapa, S.; Whyte, C.; Reed, A.W.; Jayasuriya, S.; Ye, J. Unsupervised non-rigid image distortion removal via grid deformation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2522–2532. [Google Scholar]
  4. Zhang, Z.; Tang, Y.G.; Yang, K. A two-stage restoration of distorted underwater images using compressive sensing and image registration. Adv. Manuf. 2021, 9, 273–285. [Google Scholar] [CrossRef]
  5. Tahtali, M. Imaging Techniques Through Unfavorable Media. In Propagation Through and Characterization of Atmospheric and Oceanic Phenomena; Optica Publishing Group: Vancouver, BC, Canada, 2022; p. PTh5F. 3. [Google Scholar]
  6. Jian, B.; Ma, C.; Zhu, D.; Sun, Y.; Ao, J. Seeing through Wavy Water–Air Interface: A Restoration Model for Instantaneous Images Distorted by Surface Waves. Future Internet 2022, 14, 236. [Google Scholar] [CrossRef]
  7. Treibitz, T.; Schechner, Y.; Kunz, C.; Singh, H. Flat refractive geometry. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 51–65. [Google Scholar] [CrossRef] [Green Version]
  8. Diamant, Y.; Schechner, Y.Y. Overcoming visual reverberations. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–8. [Google Scholar]
  9. Schechner, Y.Y.; Karpel, N. Attenuating natural flicker patterns. In Proceedings of the Oceans’ 04 MTS/IEEE Techno-Ocean’04 (IEEE Cat. No. 04CH37600), Kobe, Japan, 9–12 November 2004; IEEE: Piscataway, NJ, USA, 2004; Volume 3, pp. 1262–1268. [Google Scholar]
  10. Suiter, H.; Flacco, N.; Carter, P.; Tong, K.; Ries, R.; Gershenson, M. Optics near the snell angle in a water-to-air change of medium. In Proceedings of the OCEANS 2007, Vancouver, BC, Canada, 29 September–4 October 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 1–12. [Google Scholar]
  11. Lynch, D.K. Snell’s window in wavy water. Appl. Opt. 2015, 54, B8–B11. [Google Scholar] [CrossRef] [Green Version]
  12. Holst, G.C.; Krapels, K.A. Infrared Imaging Systems: Design, Analysis, Modeling, and Testing XXIV. In Proceedings of SPIE; SPIE: Bellingham, DC, USA, 2013; Volume 8706, p. 870601-1. [Google Scholar]
  13. Efros, A.; Isler, V.; Shi, J.; Visontai, M. Seeing through water. Adv. Neural Inf. Process. Syst. 2005, 7, 393–400. [Google Scholar]
  14. Donate, A.; Dahme, G.; Ribeiro, E. Classification of textures distorted by water waves. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; IEEE: Piscataway, NJ, USA, 2006; Volume 2, pp. 421–424. [Google Scholar]
  15. Wen, Z.; Lambert, A.; Fraser, D.; Li, H. Bispectral analysis and recovery of images distorted by a moving water surface. Appl. Opt. 2010, 49, 6376–6384. [Google Scholar] [CrossRef]
  16. Kanaev, A.V.; Hou, W.; Woods, S. Multi-frame underwater image restoration. In Electro-Optical and Infrared Systems: Technology and Applications VIII; SPIE: Bellingham, DC, USA, 2011; Volume 8185, pp. 208–215. [Google Scholar]
  17. Kanaev, A.V.; Hou, W.; Restaino, S.R.; Matt, S.; Gładysz, S. Restoration of images degraded by underwater turbulence using structure tensor oriented image quality (STOIQ) metric. Opt. Express 2015, 23, 17077–17090. [Google Scholar] [CrossRef]
  18. Zhang, R.; He, D.; Li, Y.; Huang, L.; Bao, X. Synthetic imaging through wavy water surface with centroid evolution. Opt. Express 2018, 26, 26009–26019. [Google Scholar] [CrossRef]
  19. Tian, Y.; Narasimhan, S.G. Seeing through water: Image restoration using model-based tracking. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 2303–2310. [Google Scholar]
  20. Tian, Y.; Narasimhan, S.G. Globally optimal estimation of nonrigid image distortion. Int. J. Comput. Vis. 2012, 98, 279–302. [Google Scholar] [CrossRef]
  21. Seemakurthy, K.; Rajagopalan, A.N. Deskewing of underwater images. IEEE Trans. Image Process. 2015, 24, 1046–1059. [Google Scholar] [CrossRef]
  22. Li, Z.; Murez, Z.; Kriegman, D.; Ramamoorthi, R.; Chandraker, M. Learning to see through turbulent water. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 512–520. [Google Scholar]
  23. James, J.G.; Agrawal, P.; Rajwade, A. Restoration of Non-rigidly Distorted Underwater Images using a Combination of Compressive Sensing and Local Polynomial Image Representations. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  24. Thapa, S.; Li, N.; Ye, J. Learning to Remove Refractive Distortions from Underwater Images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 5007–5016. [Google Scholar]
  25. Oreifej, O.; Shu, G.; Pace, T.; Shah, M. A two-stage reconstruction approach for seeing through water. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1153–1160. [Google Scholar]
  26. Hu, W.; Yuan, X.; Zhang, W.; Tan, Y. Removing Water Fluctuation via Motion Field-based Kernel Regression. J. Inf. Comput. Sci. 2014, 11, 5289–5296. [Google Scholar] [CrossRef]
  27. Halder, K.K.; Paul, M.; Tahtali, M.; Anavatti, S.G.; Murshed, M. Correction of geometrically distorted underwater images using shift map analysis. JOSA A 2017, 34, 666–673. [Google Scholar] [CrossRef]
  28. Zhang, Z.; Yang, X. Reconstruction of distorted underwater images using robust registration. Opt. Express 2019, 27, 9996–10008. [Google Scholar] [CrossRef]
  29. Sun, T.; Tang, Y.; Zhang, Z. Structural information reconstruction of distorted underwater images using image registration. Appl. Sci. 2020, 10, 5670. [Google Scholar] [CrossRef]
  30. Yang, L.; Ji, H. A variational EM framework with adaptive edge selection for blind motion deblurring [C]//. 2019: 10167-10176. Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. 2019, 10167–10176. [Google Scholar]
  31. Wen, F.; Ying, R.; Liu, Y.; Liu, P.; Truong, T.K. A simple local minimal intensity prior and an improved algorithm for blind image deblurring. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 2923–2937. [Google Scholar]
  32. Xu, Z.; Chen, H.; Li, Z. Fast blind deconvolution using a deeper sparse patch-wise maximum gradient prior. Signal Process. Image Commun. 2021, 90, 116050. [Google Scholar] [CrossRef]
  33. Aubailly, M.; Vorontsov, M.A.; Carhart, G.W.; Valley, M.T. Automated video enhancement from a stream of atmospherically-distorted images: The lucky-region fusion approach. In Atmospheric Optics: Models, Measurements, and Target-in-the-Loop Propagation III; SPIE: Bellingham, DC, USA, 2009; Volume 7463, pp. 104–113. [Google Scholar]
  34. Cox, C.; Munk, W. Slopes of the sea surface deduced from photographs of sun glitter. 1956. Available online: https://escholarship.org/uc/item/1p202179 (accessed on 28 November 2022).
  35. Cai, C.; Meng, H.; Qiao, R.; Wang, Y. Water–air imaging: Distorted image reconstruction based on a twice registration algorithm. Mach. Vis. Appl. 2021, 32, 64. [Google Scholar] [CrossRef]
  36. Alterman, M.; Swirski, Y.; Schechner, Y.Y. STELLA MARIS: Stellar marine refractive imaging sensor. In Proceedings of the 2014 IEEE International Conference on Computational Photography (ICCP), Santa Clara, CA, USA, 2–4 May 2014. [Google Scholar]
  37. Alterman, M.; Schechner, Y.Y. 3D in Natural Random Refractive Distortions; Javidi, B., Son, J.-Y., Eds.; International Society for Optics and Photonics: Bellingham, DC, USA, 2016; Volume 9867. [Google Scholar]
  38. Rauf, H.T.; Bangyal WH, K.; Lali, M.I. An adaptive hybrid differential evolution algorithm for continuous optimization and classification problems. Neural Comput. Appl. 2021, 33, 10841–10867. [Google Scholar] [CrossRef]
  39. Chang, D.; Sun, S.; Zhang, C. An accelerated linearly convergent stochastic L-BFGS algorithm. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3338–3346. [Google Scholar] [CrossRef]
  40. Rueckert, D.; Sonoda, L.; Hayes, C.; Hill, D.; Leach, M.; Hawkes, D. Nonrigid registration using free-form deformations: Application to breast MR images. IEEE Trans. Med. Imaging 1999, 18, 712–721. [Google Scholar] [CrossRef] [PubMed]
  41. Kaufman, L.; Rousseeuw, P.J. Finding Groups in Data: An Introduction to Cluster Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2009. [Google Scholar]
  42. Zhang, X.; Feng, X.; Wang, W.; Xue, W. Edge strength similarity for image quality assessment. IEEE Signal Process. Lett. 2013, 20, 319–322. [Google Scholar] [CrossRef]
  43. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. FSIM: A feature similarity index for image quality assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Ryu, E.K.; Yin, W. Large-Scale Convex Optimization: Algorithms & Analyses via Monotone Operators; Cambridge University Press: Cambridge, UK, 2022. [Google Scholar]
  46. Dai, Y.H. A perfect example for the BFGS method. Math. Program. 2013, 138, 501–530. [Google Scholar] [CrossRef]
  47. Gower, R.M.; Richtárik, P. Randomized quasi-Newton updates are linearly convergent matrix inversion algorithms. SIAM J. Matrix Anal. Appl. 2017, 38, 1380–1409. [Google Scholar] [CrossRef] [Green Version]
  48. Lin, Z.; Chen, M.; Ma, Y. The augmented lagrange multiplier method for exact recovery of corrupted low-rank matrices. arXiv preprint 2010, arXiv:1009.5055. [Google Scholar]
  49. Jian, B. Observing Through Wavy Water Surface: Reconstruction the Distortion-Free Image from a Distorted Video Sequence. 2022. (accessed on 28 November 2022). [CrossRef]
  50. Panetta, K.; Gao, C.; Agaian, S. Human-visual-system-inspired underwater image quality measures. IEEE J. Ocean. Eng. 2015, 41, 541–551. [Google Scholar] [CrossRef]
  51. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  52. He, K.; Sun, J.; Tang, X. Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1397–1409. [Google Scholar] [CrossRef]
  53. Rahaman, M.M.; Li, C.; Yao, Y.; Kulwa, F.; Rahman, M.A.; Wang, Q.; Qi, S.; Kong, F.; Zhu, X.; Zhao, X. Identification of COVID-19 samples from chest X-Ray images using deep learning: A comparison of transfer learning approaches. J. X-Ray Sci. Technol. 2020, 28, 821–839. [Google Scholar] [CrossRef]
  54. Chen, H.; Li, C.; Wang, G.; Li, X.; Rahaman, M.; Sun, H.; Hu, W.; Li, Y.; Liu, W.; Sun, C.; et al. Gashis-transformer: A multi-scale visual transformer approach for gastric histopathology image classification. arXiv e-prints 2021, arXiv:2104.14528. [Google Scholar]
  55. Ma, J.; Jiang, X.; Fan, A.; Jiang, J.; Yan, J. Image matching from handcrafted to deep features: A survey. Int. J. Comput. Vis. 2021, 129, 23–79. [Google Scholar] [CrossRef]
  56. Zhang, J.; Li, C.; Kosov, S.; Grzegorzek, M.; Shirahama, K.; Jiang, T.; Sun, C.; Li, Z.; Li, H. LCU-Net: A novel low-cost U-Net for environmental microorganism image segmentation. Pattern Recognit. 2021, 115, 107885. [Google Scholar] [CrossRef]
  57. Chen, H.; Li, C.; Li, X.; Rahaman, M.M.; Hu, W.; Li, Y.; Liu, W.; Sun, C.; Sun, H.; Huang, X. IL-MCAM: An interactive learning and multi-channel attention mechanism-based weakly supervised colorectal histopathology image classification approach. Comput. Biol. Med. 2022, 143, 105265. [Google Scholar] [CrossRef]
  58. Li, X.; Li, C.; Rahaman, M.M.; Sun, H.; Li, X.; Wu, J.; Yao, Y.; Grzegorzek, M. A comprehensive review of computer-aided whole-slide image analysis: From datasets to feature extraction, segmentation, classification and detection approaches. Artif. Intell. Rev. 2022, 55, 4809–4878. [Google Scholar] [CrossRef]
Figure 1. The framework of the image reconstruction method in this paper.
Figure 1. The framework of the image reconstruction method in this paper.
Entropy 24 01765 g001
Figure 2. The framework to reconstruct high-quality reference frame.
Figure 2. The framework to reconstruct high-quality reference frame.
Entropy 24 01765 g002
Figure 3. The recovery results of different methods. Left to right: the sample image, the results of Tian’s method [19], the results of Oreifej’s method [25], the results of Z. Zhang’ method [28], the results of T. Sun’ method [29] and the result of our method.
Figure 3. The recovery results of different methods. Left to right: the sample image, the results of Tian’s method [19], the results of Oreifej’s method [25], the results of Z. Zhang’ method [28], the results of T. Sun’ method [29] and the result of our method.
Entropy 24 01765 g003
Figure 4. The running time of different methods after 5 iterations. (a) A comparison of total running time of different methods; (b) A comparison of running time of different methods after each iteration.
Figure 4. The running time of different methods after 5 iterations. (a) A comparison of total running time of different methods; (b) A comparison of running time of different methods after each iteration.
Entropy 24 01765 g004
Figure 5. The running times of different methods at the same precision threshold. (a) The total running time of each method when the criterion for termination is satisfied; (b) The number of iterations for each method when the criterion for termination is satisfied.
Figure 5. The running times of different methods at the same precision threshold. (a) The total running time of each method when the criterion for termination is satisfied; (b) The number of iterations for each method when the criterion for termination is satisfied.
Entropy 24 01765 g005
Figure 6. The restoration results with different methods in water-to-air imaging scenarios.
Figure 6. The restoration results with different methods in water-to-air imaging scenarios.
Entropy 24 01765 g006
Figure 7. The fused frame based on the different image quality metrics at the 2th iteration stage. From left to right: ESSIM, SSIM, FSIM and NIQE.
Figure 7. The fused frame based on the different image quality metrics at the 2th iteration stage. From left to right: ESSIM, SSIM, FSIM and NIQE.
Entropy 24 01765 g007
Figure 8. Evolution of the reference frame of the different methods in registration process. Top to bottom: the ground-truth image, the result of Oreifej’s method [25], the results of Z. Zhang’s method [28], the results of T. Sun’s method [29] and the results of our method.
Figure 8. Evolution of the reference frame of the different methods in registration process. Top to bottom: the ground-truth image, the result of Oreifej’s method [25], the results of Z. Zhang’s method [28], the results of T. Sun’s method [29] and the results of our method.
Entropy 24 01765 g008
Figure 9. Evolution of the synthetic frame I F in registration process. Top to bottom: Our method employed without blurring or deblurring, with reference frame deblurring using [52], and with frame blurring using our method.
Figure 9. Evolution of the synthetic frame I F in registration process. Top to bottom: Our method employed without blurring or deblurring, with reference frame deblurring using [52], and with frame blurring using our method.
Entropy 24 01765 g009
Figure 10. The comparison of running time of different optimization strategy after each iteration.
Figure 10. The comparison of running time of different optimization strategy after each iteration.
Entropy 24 01765 g010
Figure 11. The comparison between James’s method [23] and the proposed method. Left to right: the sample images, the results of James’s [23], and the results of our method.
Figure 11. The comparison between James’s method [23] and the proposed method. Left to right: the sample images, the results of James’s [23], and the results of our method.
Entropy 24 01765 g011
Table 1. The frame size and patch size of the different video sequences.
Table 1. The frame size and patch size of the different video sequences.
BrickCheckerboardLarge FontsMiddle FontsSmall Fonts
frame size 268 × 292 238 × 285 238 × 324 253 × 293 255 × 284
patch size 76 × 116 68 × 114 68 × 108 84 × 58 102 × 94
Table 2. Numerical comparison based on the standard data sets 1.
Table 2. Numerical comparison based on the standard data sets 1.
Data SetsMethodsMSE(L)PSNR(H)SSIM(H)UIQM(H)
Middle fontsTian [19]0.007921.02750.65982.5215
Oreifej [25]0.007321.38290.74122.6182
Z. Zhang [28]0.016417.84070.57732.5773
T. Sun [29]0.011919.23610.61372.6277
Our method0.005322.75910.78912.7056
Small fontsTian [19]0.004323.69160.64752.5852
Oreifej [25]0.003824.14900.70982.8036
Z. Zhang [28]0.006122.17680.56892.7409
T. Sun [29]0.006821.70690.54242.7853
Our method0.003624.47520.72432.8409
BricksTian [19]0.010819.68000.60172.4071
Oreifej [25]0.005122.92780.63612.5631
Z. Zhang [28]0.018217.39510.41122.4807
T. Sun [29]0.011819.27000.43732.5846
Our method0.005122.92780.64972.7126
Large fontsTian [19] × × × 2.2026
Oreifej [25] × × × 2.0533
Z. Zhang [28] × × × 2.1072
T. Sun [29] × × × 2.1218
Our method × × × 2.2851
CheckerboardTian [19] × × × 2.4701
Oreifej [25] × × × 2.4900
Z. Zhang [28] × × × 2.5504
T. Sun [29] × × × 2.5240
Our method × × × 2.5928
1 Because the template images of “Large ” and “Checkerboard” are unavailable, the PSNR, MSE, SSIM are presented with “ ×
Table 3. Numerical comparison based on Zhang’s data sets [18] (Bold values are the best).
Table 3. Numerical comparison based on Zhang’s data sets [18] (Bold values are the best).
Data SetsUIQM(H)
Oreifej/Z. Zhang/T. Sun/Our method
ripple1.9231/1.9544/2.0128/2.1573
gravity1.3119/1.3218/1.3266/1.3347
Table 4. Image quality evaluation results of reference frames for different methods.
Table 4. Image quality evaluation results of reference frames for different methods.
IterationMethodsMSE(L)PSNR(H)SSIM(H)
1thOreifej0.012019.19340.4847
Z. Zhang0.014918.26340.3940
T. Sun0.011719.30800.4907
Our method0.012019.19340.4847
2thOreifej0.010719.72440.5576
Z. Zhang0.017117.66890.4293
T. Sun0.012219.14310.5904
Our method0.007221.42570.6491
3thOreifej0.007321.39320.7123
Z. Zhang0.022516.47420.4364
T. Sun’s0.013018.86770.6054
Our method0.005722.43230.7475
4thOreifej0.007221.42470.7340
Z. Zhang0.012119.17180.6115
T. Sun0.013218.79930.6086
Our method0.005622.54410.7700
5thOreifej0.007221.42470.7407
Z. Zhang0.017217.65360.5737
T. Sun0.013318.77260.6090
Our method0.005522.59860.7781
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jian, B.; Ma, C.; Zhu, D.; Huang, Q.; Ao, J. Water-Air Interface Imaging: Recovering the Images Distorted by Surface Waves via an Efficient Registration Algorithm. Entropy 2022, 24, 1765. https://doi.org/10.3390/e24121765

AMA Style

Jian B, Ma C, Zhu D, Huang Q, Ao J. Water-Air Interface Imaging: Recovering the Images Distorted by Surface Waves via an Efficient Registration Algorithm. Entropy. 2022; 24(12):1765. https://doi.org/10.3390/e24121765

Chicago/Turabian Style

Jian, Bijian, Chunbo Ma, Dejian Zhu, Qihong Huang, and Jun Ao. 2022. "Water-Air Interface Imaging: Recovering the Images Distorted by Surface Waves via an Efficient Registration Algorithm" Entropy 24, no. 12: 1765. https://doi.org/10.3390/e24121765

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop