Next Article in Journal
Disturbance Rejection and Control Design of MVDC Converter with Evaluation of Power Loss and Efficiency Comparison of SiC and Si Based Power Devices
Next Article in Special Issue
Traffic Police Gesture Recognition Based on Gesture Skeleton Extractor and Multichannel Dilated Graph Convolution Network
Previous Article in Journal
Bus Dynamic Travel Time Prediction: Using a Deep Feature Extraction Framework Based on RNN and DNN
Previous Article in Special Issue
Robust Image Classification with Cognitive-Driven Color Priors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Overwater Image Dehazing via Cycle-Consistent Generative Adversarial Network

School of Computer Science and Technology, Harbin Institute of Technology, Weihai 264209, China
*
Author to whom correspondence should be addressed.
Electronics 2020, 9(11), 1877; https://doi.org/10.3390/electronics9111877
Submission received: 30 September 2020 / Revised: 1 November 2020 / Accepted: 4 November 2020 / Published: 8 November 2020

Abstract

:
In contrast to images taken on land scenes, images taken over water are more prone to degradation due to the influence of the haze. However, existing image dehazing methods are mainly developed for land-scene images and perform poorly when applied to overwater images. To address this problem, we collect the first overwater image dehazing dataset and propose a Generative Adversial Network (GAN)-based method called OverWater Image Dehazing GAN (OWI-DehazeGAN). Due to the difficulties of collecting paired hazy and clean images, the dataset contains unpaired hazy and clean images taken over water. The proposed OWI-DehazeGAN is composed of an encoder–decoder framework, supervised by a forward-backward translation consistency loss for self-supervision and a perceptual loss for content preservation. In addition to qualitative evaluation, we design an image quality assessment neural network to rank the dehazed images. Experimental results on both real and synthetic test data demonstrate that the proposed method performs superiorly against several state-of-the-art land dehazing methods. Compared with the state-of-the-art, our method gains a significant improvement by 1.94% for SSIM, 7.13% for PSNR and 4.00% for CIEDE2000 on the synthetic test dataset.

1. Introduction

Images of overwater scenes play an important role in human image galleries. However, these images are prone to degradation due to thick mist that are often appearing over lakes, rivers, and seas. Although numerous image dehazing methods have been developed [1,2,3,4,5], our experiments show that these methods perform far from satisfying since they are originally designed for land scene images, of which the data distribution differs significantly.
Hazy images are usually modeled as I ( x ) = J ( x ) t ( x ) + A ( 1 t ( x ) ) , where I ( x ) and J ( x ) are the observed hazy image and the scene, respectively [6,7]. The symbol x denotes a pixel index, and A is the global atmospheric light. t ( · ) denotes the transmission map, which describes the portion of light that is not scattered and reaches the camera sensors. When the haze is homogeneous, t ( · ) can be defined as: t ( x ) = e β d ( x ) , where β is the scattering coefficient and d ( x ) is the distance between objects and the camera.
Existing methods fall into two categories according to the type of features they used: methods based on hand-crafted features [1,2,8,9,10,11,12] or methods based on convolutional neural network (CNN) features [3,4,5,13,14,15,16]. The former generally focuses on estimating the global atmospheric light intensity A ( · ) and the transmission map t ( · ) , and hence their performance are susceptible to estimation errors of A ( · ) or t ( · ) . To alleviate these limitations, the latter, which is based on CNNs [17] or Generative Adversial Networks (GANs) [18], aims to directly estimate clean images in a data-driven scheme. Although promising dehazing results have been achieved, existing CNN- or GAN- based methods perform not well on overwater images as shown in Figure 1b,c. This is because they are designed for dehazing land scene images, which results from the difference among data distributions of land images and overwater images. Another issue is that existing image dehazing datasets [12,19,20,21,22] are dominated by land scenes, which have significant data distribution differences compared to that of overwater images.
In this paper, we address both the above-mentioned issues. First, we construct a new dataset, named HazyWater, especially for dehazing overwater images. Since collecting paired hazy and clear images is difficult and expensive, the HazyWater dataset is composed of unpaired hazy and clean overwater images. Although the unpaired character challenges most of the existing methods, we show that satisfying performance could be achieved by the proposed dehazing method in the experiment section. Second, we propose an OverWater Image Dehazing GAN (OWI-DehazeGAN) inspired by CycleGAN [23] to directly recover clean images. OWI-DehazeGAN employs the forward-backward translation consistency loss as self-supervision to tackle the unpairness of training data. We also introduce a perceptual loss to enhance the recovered image quality.
Our contributions are summarized as follows:
  • We create the first overwater image dehazing dataset. Most of the existing image dehazing methods (including CNN-based methods) perform not well on overwater images, and we hope this dataset is able to facilitate the research in this field.
  • We propose an OWI-DehazeGAN to dehaze overwater images, which is based on but performs superior to Cycle-GAN. The proposed network is able to utilize unpaired training data and preserve image details simultaneously.
  • We propose an image quality assessment network to rank the dehazed images of dehazed methods, which facilitates the comparison of different dehazing methods. Extensive experimental results evaluated by this network as well as the peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and CIEDE2000 metrics demonstrate the effectiveness of the proposed OWI-DehazeGAN.

2. Related Work

2.1. Methods Based on Hand-Crafted Features

Many efforts have been devoted to image dehazing in the past decades based on hand-crafted features [1,2,8,9,10,11,12]. Tan et al. [8] proposed a contrast maximizing approach using Markov random fields (MRF) based on the observation that clean images have higher contrast than hazy ones. In Reference [9] Tarel et al. propose a fast dehazing method by combining atmospheric veil inference, image restoration and smoothing tone mapping. Later, He et al. [10] estimated the transmission map by utilizing dark-channel prior (DCP). Meng et al. [11] explores the inherent boundary constraint on the transmission function. In order to recover depth information, Zhu et al. [1] proposed a color attenuation prior (CAP) by creating a linear model on local priors. Different from previous methods that use various patch-based priors, Berman et al. [2] presented a new image dehazing algorithm based on non-local prior so that a haze-free image is able to be well approximated by a few distinct colors.
While the afore-mentioned methods have achieved promising results, they perform far from satisfying when applied to overwater images. For example, MRF [8] tends to produce over-saturated images. The enhanced images of FVR [9] often contain distorted colors and severe halos. DCP [10] does not work well when it comes to the sky regions, as the scene objects are similar to the atmospheric light. This is mainly because the major region of overwater images are water and sky, which have a different data distribution from that of land scene images.

2.2. Methods Based on CNN Features

Deep convolutional neural networks have shown promising success in various computer vision tasks [24,25]. Many CNN-based image dehazing methods have been proposed. Cai et al. [13] propose an end-to-end DehazeNet with non-linear regression layers to estimate medium transmission. Instead of estimating the transmission map or the atmospheric light firstly, AOD-Net [16] predicts the haze-free images directly using a light-weight CNN. Proximal Dehaze-Net [4] combines the advantages of traditional prior-based dehazing methods and deep learning methods by incorporating haze-related prior learning into the deep network.
Since Goodfellow [18] proposed the GAN method in 2014, there have been many effective variants tailored to different computer vision tasks [23,26,27]. Motivated by the success of GANs in style transfer [28], super-resolution [29], text to image [30], and image inpainting [31], many GAN-based methods have been proposed for image dehazing. In Reference [5], a Densely Connected Pyramid Dehazing Network (DCPDN) is proposed to jointly learn the transmission map, atmospheric light and dehazing result all together. Yang et al. [32] propose to loose the paired training constraint by introducing a disentanglement and reconstruction mechanism. Li et al. [14] designed a solution based on a cGAN network [26] to directly estimate the clean image. Ren et al. [3] adopt an ensemble strategy to take advantage of the information in white balance, contrast-enhancing, and gamma correction images. Overall, these methods are trained on paired data, which is unsuitable for the proposed overwater image dehazing task, where only unpaired training data is available.

2.3. Image Dehazing Dataset

Image dehazing tasks profit from the continuous efforts for large-scale data. Several datasets [12,19,20,21,22] have been introduced for image dehazing. Fattal et al. [12] provide 12 high-quality synthetic images for outdoor image dehazing. For evaluating the performance of automatic driving systems in various hazy conditions, FRIDA [33] introduces a dataset that includes synthetic images. MSCNN [15] and AOD-Net [16] utilize the indoor NYU2 Depth Database [34] and the Middlebury Stereo database [35] to synthesize hazy images using the known depth information. O-HAZE [20] is an outdoor scenes dataset, which is composed of pairs of real hazy and corresponding clean data. I-HAZE [20] is a dataset that contains 35 image pairs of hazy and corresponding ground-truth indoor images. Foggy Cityscapes dataset [21] is a synthetic version of the Cityscapes dataset using incomplete depth information. Li [22] launched a new large-scale benchmark which is made up of synthetic and real-world hazy images, called Realistic Single Image Dehazing (RESIDE). However, most datasets are synthetic and not tailored to the problem of overwater image dehazing. Different from the above datasets, we collect a dataset that contains real data especially used for dehazing overwater images.

3. Proposed Method

In this section, we first introduce the overall structure of the proposed OWI-DehazeGAN, and then detail each of its components. At the end of this section, we present the proposed image quality ranking network, which is later used to rank the outputs of different dehazing methods in the experiment section.
Figure 2 shows the main architecture of the proposed OWI-DehazeGAN. Unlike traditional GANs, OWI-DehazeGAN consists of two generators (G and F) and two discriminators ( D x and D y ) in order to be trainable with unpaired training data. Specifically, generator G predicts clean images Y from hazy images X, and F vice versa. D x and D y distinguish hazy images and clean images, respectively. Below we provide more details about each component.

3.1. Generator

We adopt the same structure for the two generators G and F. Both generators are divided into three parts: encoding, transformation, and decoding. The architecture and details of the generator are shown Figure 3a and Table 1, respectively.
Encoding: The encoding module extracts image features by three convolution layers, which serve as down-sampling layers to decrease the resolution of the original input. Each convolution layer is followed by an instance normalization [36] and Leaky Relu [37]. Since image dehazing can be treated as a domain adaptation problem that views each image as a domain, instance normalization is more suitable for image dehazing than batch normalization [38]. Leaky Relu is an improved activation method for Relu [39] which has all the benefits of Relu and solves the dead Relu problem.
Transformation: The transformation module translates information from one domain to another via nine ResNet blocks [40]. ResNet block in our network contains two 3 × 3 convolution layers with the same number of filters. Due to the results of image dehazing, the need to retain the characteristics of the original image, such as the shape and color, the ResNet block is well-suited to accomplish these transformations.
Decoding: The decoding module includes the up-sampling operations and nonlinear mappings. There are several choices for upsampling, such as deconvolution [41], sub-pixel convolution [42] and resize convolution [43]. In order to reduce checkerboard artifacts [43] caused by deconvolution or sub-pixel convolution, we use the resize convolution for decoding. It increases the resolution of the feature map using nearest-neighbor interpolation followed by convolution. Inspired by the success of U-Net [44], we introduce two symmetric skip connections to deliver information between encoding and decoding modules. Finally, images are recovered through convolution and tanh activation.

3.2. Discriminator

We use two discriminators D x and D y to distinguish the input hazy images and clean images, respectively. The discriminator is implemented in a fully convolutional fashion. We use four convolution blocks in discriminator. The first block consists of convolution and Leaky Relu, the last block only contains convolution and the remaining blocks are composed of convolution, instance normalization and Leaky Relu. In order to maintain the stability of GAN training process, we do not utilize the sigmoid activation in the last block. The architecture and details of the discriminator are shown in Figure 3b and Table 2.

3.3. Loss Function

We utilize three kinds of losses to train the proposed network, including Adversarial Loss, Cycle Consistency Loss, and Perceptual Loss. The Adversarial Loss and the Cycle Consistency Loss enable the proposed network trainable with unpaired data, and the Perceptual Loss preserves image details.

3.3.1. Adversarial Loss

As done in CycleGAN, we use adversarial loss and cycle consistency loss for unpaired training data. x X , y Y are hazy image and clean image unpaired data, respectively. For the generator G and discriminator D y , the adversarial loss is formulated as:
L G A N ( G , D y , x , y ) = l o g ( D y ( y ) ) + l o g ( 1 D y ( G ( x ) ) ) .
Correspondingly, the constraint on generator F and its discriminator D x is
L G A N ( F , D x , x , y ) = l o g ( D x ( x ) ) + l o g ( 1 D x ( F ( y ) ) ) .
However, the above losses are prone to unstable training and generating low quality images. To avoid the vanishing gradients problem and achieve high quality images, we use a least squares loss [27] instead of the negative log likelihood objective [18]. Therefore, Equations (1) and (2) are modified as:
L a d v ( G , D y , x , y ) = 1 2 [ ( D y ( G ( x ) ) 1 ) 2 + ( D y ( y ) 1 ) 2 + D y ( G ( x ) ) 2 ]
L a d v ( F , D x , x , y ) = 1 2 [ ( D x ( F ( y ) ) 1 ) 2 + ( D x ( x ) 1 ) 2 + D x ( F ( y ) ) 2 ] .
The final adversarial loss is denoted as:
L a d v ( G , F , D x , D y ) = L a d v ( G , D y , x , y ) + L a d v ( F , D x , x , y ) .

3.3.2. Cycle Consistency Loss

CycleGAN introduces a cycle consistency loss to solve the problem that an adversarial loss alone cannot guarantee that the output distribution matches the target distribution. For each image x, F ( G ( x ) ) is able to bring G ( x ) back to the original image. Similarly, G ( F ( y ) ) is able to bring F ( y ) back to the original image y. F ( G ( x ) ) is the cyclic image of input x, and G ( F ( y ) ) is the cyclic image of the original image y. To train generators G and F at the same time, the consistency loss includes two constraints: F ( G ( x ) ) x , G ( F ( y ) ) y . Cycle consistency loss is defined to calculate L1-norm between the input and cyclic image for unpaired image dehazing:
L c y c ( G , F ) = | | F ( G ( x ) ) x | | 1 + | | G ( F ( y ) ) y | | 1 .

3.3.3. Perceptual Loss

Inspired by the success of perceptual loss in style transfer [45], we introduce perceptual loss to restrict the reconstruction of image details. Instead of measuring per-pixel difference between the images, perceptual loss is concerned with the distinction between feature maps, which comprises various aspects of content and perceptual quality. The perceptual loss is defined as:
L p e r ( G , F ) = | | θ ( x ) θ ( F ( G ( x ) ) ) | | 2 2 + | | θ ( y ) θ ( G ( F ( y ) ) ) | | 2 2 .
Here, θ represents the feature maps generated from the relu4_2 on pertained VGG-16 [46] network.

3.3.4. Objective Function

Our final loss is defined as the weighted sum of previous losses:
L ( G , F , D x , D y ) = L a d v ( G , F , D x , D y ) + λ 1 L c y c ( G , F ) + λ 2 L p e r ( G , F ) ,
where coefficients λ 1 and λ 2 represent the weights of cycle consistency loss and perceptual loss, respectively. We found that giving an over-weight to perceptual loss causes the instability of training process, which could make the GAN model non-convergent, thus the weight of perceptual loss is much less than the weight of cyclic consistency loss. In training process, we minimize the generators G, F and maximize the discriminators D x , D y .
The final objective function is:
< G * , F * > = arg min G , F max D x , D y L ( G , F , D x , D y ) .
We summarize the training procedure in Algorithm 1.
Algorithm 1 OWI-DehazeGAN training procedure pseudocode.
Input: The hazy training dataset X; The clean training dataset Y; Training epoch number e p o c h .
Output: The well-trained generators G and F; The well-trained discriminators D x and D y .
1:
for  e p o c h = 1 ; e p o c h e p o c h m a x ; e p o c h + + do
2:
   Draw mini-batches of samples { x ( 1 ) , ⋯, x ( m ) } from X;
3:
   Draw mini-batches of samples { y ( 1 ) , ⋯, y ( m ) } from Y;
4:
   for each mini-batch do
5:
    Compute the discriminator loss on real images;
6:
    Compute the discriminator loss on fake images;
7:
    Update the discriminators D x and D y ;
8:
   end for
9:
   for each mini-batch do
10:
    Compute X G ( x ) F ( G ( x ) ) and Y F ( Y ) G ( F ( Y ) ) generator loss in Equation (6);
11:
    Compute the perceptual loss in Equation (7);
12:
    Update the generators G and F;
13:
   end for
14:
end for

3.4. Dehazed Image Quality Assessment

In order to verify the effectiveness of the proposed OWI-DehazeGAN, we design a dehazed image quality assessment model based on natural image statistics and the VGG network. Natural image refers to the image directly obtained from the natural scene by using optical photographic instruments such as cameras. Natural images are different from the distorted image. Natural images are directly captured from natural scenes, so they have some natural properties. By making statistics on these properties, natural scene statistics (NSS [47]) of images can be obtained. Due to the differences between natural images and distorted images in NSS, NSS has been widely used in image quality assessment, especially no-reference image quality assessment.
In this paper, the NSS we use is mean substracted contrast normalization (MSCN [48]) coefficients, which is used to normalize a hazy image. After normalization pixel intensities of haze-free by MSCN coefficient follow a Gaussian Distribution while pixel intensities of hazy images do not. The deviation of the distribution from an ideal bell curve is therefore a measure of the amount of distortion in the image. To calculate the MSCN Coefficients, the image intensity I ( i , j ) at pixel ( i , j ) is transformed to the luminance I ^ ( i , j ) . I ^ ( i , j ) is defined as:
I ^ ( i , j ) = I ( i , j ) μ ( i , j ) σ ( i , j ) + C ,
where μ ( i , j ) and σ ( i , j ) represent the local mean field and local variance field obtained by calculating the image using a gaussian window with a specific size. Local Mean Field μ is the Gaussian Blur of the input image. Local Variance Field σ is the Gaussian Blur of the square of the difference of original image and μ . C is a constant, in case the denominator is zero. The calculation of MSCN coefficient is shown in Figure 4. When dehazed image is normalized by MSCN coefficient, only uniform appearance and edge information are retained. Human eyes are very sensitive to edge information, so the normalized image is consistent with human vision.
The proposed image quality assessment (IQA) model for dehazed images is divided into three parts: luminance normalization, feature extraction and regression of evaluation score. The dehazed images are firstly normalized by the MSCN coefficient, which provides a good normalization of image luminance and does not have a strong dependence on the intensity of texture. Then, taking the normalized image into the convolution layers of VGG-16 to extract features, and finally predicting an image quality score between 0 and 9 through two fully connection layers. The units of two fully connection layers are 512 and 1, respectively. The architecture of the IQA model is shown in Figure 5.
The loss function of this IQA model is MAE. The loss is defined as:
l o s s I Q A = 1 N i = 1 n | y i y i * | ,
where N represents the number of images in the training set, y i and y i * denote target data and output data, respectively. The optimization goal of the IQA model in the training phase is to minimize the average absolute error loss. Learning the mapping between dehazed images and corresponding Mean Opinion Scores (MOS [49]) is achieved by minimizing the loss between the predicted score y i * and the corresponding ground truth y i .

4. Experiment

In this section, we first present our HazyWater dataset. Then we quantitatively evaluate the OWI-DehazeGAN on a synthetic dataset and real-world hazy data, with comparisons to several state-of-the-art methods. Experimental details are also explained in this section.

4.1. Dataset

We collect a real unpaired image dataset called HazyWater Dataset for the image dehazing in overwater scenes. All the data were gathered from Google. The training set consists of 4531 unpaired images, which are 2090 hazy images and 2441 clean images. These training images were all resized to 640 × 480 . Figure 6 illustrates some examples of our dataset. There were three main differences between the proposed dataset and the existing datasets: (1) the HazyWater dataset is a large-scale natural real dataset with hazy images and unpaired haze-free images, while the previous datasets are only composed of synthetic data; (2) the HazyWater dataset is tailored to the task of overwater image dehazing, rather than focusing on indoor or outdoor scenes; (3) the proposed dataset is much more challenging because the regions of sky and water surface make up a large part of the image.
In order to evaluate different image dehazing methods in overwater scene subjectively, we intorduced a natural overwater testing, which contains 127 challenging hazy images collected from overwater scenes. To quantitatively compare different image-dehazing methods, we selected 90 (30 images × 3 medium extinction coefficients β ) overwater hazy images with corresponding ground-truth from the RESIDE OTS dataset [22]. RESIDE OTS dataset is a large scale synthetic dataset in outdoor scene with a handful of overwater images. We apply SSIM [50], PSNR [51] and CIEDE2000 [52] to the dehazed results on synthetic images. Based on the HazyWater Dataset, we compared our proposed method against several state-of-the-art dehazing methods in real and synthetic data, including: DCP [10], FVR [9], BCCR [11], CAP [1], DehazeNet [13], MSCNN [15], AOD-Net [16], dehaze-cGAN [14], Proximal Dehaze-Net [4], CycleGAN [23].

4.2. Experimental Settings

The input images of generators and discriminators were set to 256 × 256 during training. We used an Adam [53] solver to optimize gradient with a learning rate of 2 × 10 4 . The batch size was 1. The weight of cyclic consistency loss λ 1 and perceptual loss λ 2 were 10 and 0.0001, respectively. The coefficient α of Leaky Relu was 0.2. The update proportion was 1 for generators G, F and discriminators D x , D y . The proposed OWI-DehazeGAN was implemented in TensorFlow with a Nvidia GTX 1080 Ti GPU.

4.3. Qualitative Results on Real Images

Figure 7 and Figure 8 show several dehazing results of the proposed algorithm against the state-of-the-art methods. DCP [10] tends to overestimate the thickness of the haze and produce dark results (Figure 7b and Figure 8b). The dehazed images by FVR [9] and BCCR [11] have significant color distortions and miss most details as shown in Figure 7c,d and Figure 8c,d. The best performer in the hand-crafted prior based methods was CAP [1], which generally reconstructs details of haze-free images. The deep learning-based approach achieved comparable results, such as DehazeNet [13], MSCNN [15] and dehaze-cGAN [14]. However, these results indicate that existing methods cannot handle overwater hazy images very well. For example, the dehazed results (Figure 7f,g and Figure 8f,g) by MSCNN and DehazeNet have a similar problem that tends to magnify the phenomenon of color cast and have some remaining haze. The illumination of the Proximal Dehaze-Net [4] and AOD-Net [16] results is dark, as shown in Figure 7h,j and Figure 8h,j, which are not consistent with human visual perception. From Figure 7k, CycleGAN generates a lot of pseudo-colors in heavy fog conditions, which is quite different from the original colors. Meanwhile, the results of CycleGAN generate extensive checkerboard artifacts in the sky regions. In contrast, the dehazed results by our method shown in Figure 7l and Figure 8l are visually pleasing in heavy hazy or light mist condition.

4.4. Qualitative and Quantitative Results on Synthetic Images

We further conducted some experiments based on synthetic hazy images. Although the proposed method is trained on real unpaired data, we note that it can be applied for synthetic images as well. Figure 9 shows some dehazed images generated by various methods. Figure 9a shows the groundtruth as reference. As shown in Figure 9b–d, the results of DCP [10], FVR [9] and BCCR [11] have some color or detail distortion. The dehazed results by CAP [1] (Figure 9e), DehazeNet [13] (Figure 9f), MSCNN [15] (Figure 9g), AOD-Net [16] (Figure 9h), dehaze-cGAN [14] (Figure 9i) and Proximal Dehaze-Net [4] (Figure 9j) are closer to groundtruth Figure 9a than the results based on priors. However, there was still some remaining haze as shown in Figure 9e–h. The result generated by CycleGAN in Figure 9k shows that there exists serious color cast and losses of some color information. The dehazed result generated by our approach in Figure 9l, by contrast, is visually close to the groundtruth image.
An advantage of testing on synthetic data is able to objectively evaluate experimental results via SSIM, PSNR and CIEDE2000. SSIM, PSNR and CIEDE2000 provide a pixel-wise measure between clean images and dehazed images. A higher SSIM score indicates that the generated results are more consistent with human perception. PSNR forecasts the effectiveness of image dehazing, and CIEDE2000 presents that smaller scores indicate better color preservation. In Figure 9, the SSIM and PSNR values also indicate that our method surpasses other methods. From Table 3, our method get higher PSNR, higher SSIM and lowerCIEDE2000 on the synthetic dataset. Compared with the state-of-the-art, our method gained a significant improvement by 1.94% for SSIM, 7.13% for PSNR and 4.00% for CIEDE2000, and significantly better than CycleGAN. In addition, the generator F of our proposed method can be used to generate paired image dehazing data, as shown in Figure 10. In general, the dehazed results by the proposed algorithm had higher visual quality and fewer color distortions.

4.5. Dehazed Image Ranking

The proposed IQA model for dehazed images was pre-trained on TID2013 [49] and then fine-tuned using the IVC Dehazing Dataset [19]. The TID2013 includes different types of image distortion, while IVC Dehazing Dataset is designed to evaluate the quality of dehazed images. Predicted scores were used to qualitatively rank photos as shown in Figure 11. Ranking scores and the rankings are presented below each image, where ‘1’ denotes the best visual perception and ‘10’ for the worst. Figure 11 shows that the quality of overwater dehazed images generated by OWI-DehazeGAN is better than other methods. For a comprehensive comparison, we also report the dehazed image quality measured by four typical image quality assessment methods in Table 4. The best result is shown in red font. Table 4 shows that the proposed method achieves the best performance in terms of almost all metrics.

4.6. Run Time

We also conducted the running speed experiment to evaluate the efficiency of each method. For fairness, we ran all methods on the same platform (Intel Core i7-7800X CPU@3.50 GHz, 32 GB memory). In the experiment, 90 synthetic images were used for testing, each of which were resized to 512 × 512 pixels. DCP [10], FVR [9] and CAP [1] are run without GPU acceleration. DehazeNet [13], MSCNN [15], AOD-Net [16], dehaze-cGAN [14] and our model run with a GPU card (NVIDIA GeForce GTX 1080 Ti). Table 5 shows more implementation details and the average run time. AOD-Net [16] ran fastest thanks to its lightweight structure. Our method ran slightly slower than AOD-Net (0.07 s behind) but ran 0.34 s faster than the third best method. Overall, the proposed method performed comparably against state-of-the-art methods in terms of the running speed.

5. Analysis and Discussion

In this section, we further explore and analyze the effect of each component of the proposed method. Firstly, we explore the robustness of the proposed method in image color cast removal. Then, we discuss the advantage of resizing convolution comparing to other up-sampleing methods. We also analyze the effect of perceptual loss for image the dehazing task. Finally, we show the limitation of the proposed algorithm.

5.1. Robustness to Color Cast

In the task of image dehazing, if the global atmospheric light is estimated incorrectly, it is easy to cause image color cast. The proposed OWI-DehazeGAN is robust to color cast removal. Existing methods, such as DehazeNet and MSCNN, will amplify this disadvantage and reduce image quality. An effective measure to solve the image color cast problem is color balance. The purpose of color balance is to make sure that the intrinsic color of the objects in the scene does not change under different illumination conditions. Our method plays an important role in color balance which can remove the color cast. As shown in Figure 12, our method performed well when the image encounters color cast. Compared to DCP, DehazeNet and MSCNN, our method effectively corrects the color cast phenomenon and produces a better visual image. Although our method cannot produce completely haze-free results with the best visual quality, while properly retaining a certain haze is beneficial to perceiving image depth.

5.2. Effect of Resize Convolution

In the decoding process of the generator, we use resize convolutions to increase the resolution of the feature maps, rather than deconvolution or sub-pixel convolution. To better understand how the resize convolution contributes to our proposed method, we train three end-to-end networks with different upsample mode: (i) deconvolution, (ii) sub-pixel convolution, and (iii) resize convolution.
Figure 13 shows the results of the deconvolution, sub-pixel convolution, and resize convolution as the upsampling mode in our network. In comparison, the results of resize convolution are best viewed from the perspective of the human perception and retain more detailed information. It can be seen from Figure 13b, that there are many artifacts in the regions of sky and water surface. From Figure 13f which is the zoom-in views in Figure 13b, we observe plenty checkerboard pattern of artifacts caused by deconvolution. Although the sub-pixel convolution (Figure 13c) can alleviate the “checkerboard artifacts” to some extent, the result of sub-pixel convolution is rough and unsatisfying. Comparing with the first two approaches, resize convolution recovers most scene details and maintain the original colors. Furthermore, it is obvious that the result of resize convolution is the best from the comparison of the zoom-in views. From Table 6, the introduced resize convolution gains higher PSNR, SSIM scores and a lower CIEDE2000 score than deconvolution and sub-pixel convolution, which indicate that resize convolution can generate visually perceptible images.

5.3. Effect of Perceptual Loss

In this section, we analyze how the introduced perceptual loss helps estimate high-quality haze-free images. To show the effectiveness of our loss function, we also train an overwater image dehazing network without perceptual loss. We show an example of result with and without perceptual loss as shown in Figure 14, the generated direction for images in the first row is X G ( X ) F ( G ( X ) ) , the second row is the opposite of the first row ( Y F ( Y ) G ( F ( Y ) ) ). We observed from Figure 14c,d,g,h that the estimated haze-free images and cyclic images lack fine details and the regions of the sky do not match with the input hazy image, which leads to the dehazed results containing halo artifacts when the perceptual loss is not used. Through the comparison of Figure 14b,d we can also find that the perceptual loss is favorable for the reconstruction of the sky regions, which is very necessary for the overwater image dehazing. Meanwhile, comparing Figure 14c,h, we can know that the perceptual loss has a little inhibition to the colour cast.
From Table 7, we observe that results generated by our network with perceptual loss gain higher PSNR, SSIM and lower CIEDE2000 score. Higher SSIM and PSNR scores suggest that the proposed method with perceptual loss has a better image dehazing effect and is consistent with human perception. Lower CIEDE2000 means less color difference between dehazed image and groundtruth. The above experiments show that the proposed loss is effective for overwater image dehazing.

5.4. Limitation

Although the proposed algorithm is effective in most cases, a limitation of our method is that it cannot well handle degraded overwater images with very heavy haze, which are close to pure white. As shown in Figure 15, our network fail to adopt to this situation because the extremely heavy haze causes the illumination of water and sky regions to be close to atmospheric light.

6. Conclusions

In this paper, we formulate an overwater image dehazing task, create the first overwater image dehazing dataset, and propose the OWI-DehazeGAN to dehaze overwater images. Compared to previous CNN-based methods which require paired training data, the proposed OWI-DehazeGAN is able to be trained unpaired images. Our method directly predicts clean images from hazy input bypassing to estimate transmission maps and global atmospheric lights. We utilize the perceptual loss and resize convolution to preserve detailed textures and to alleviate the checkerboard artifacts. Extensive experimental results demonstrate that the presented method produces superior results than most state-of-the-art dehazing methods.

Author Contributions

S.Z., J.S. and Q.L. were responsible for the overall design of the study. S.Z. performed the experiments and drafted the manuscript. Y.Q. and J.Y. revised the manuscript. All authors read and agreed to the published version of the manuscript.

Funding

This work was supported in part by National Science and Technology Major Project under Grant No.2018YFB0804703, National Natural Science Foundation of China (Nos. 61902092 and 61872112), Fundamental Research Funds for the Central Universities Grant No.HIT.NSRIF.2020005, and National Key Research and Development Program of China (Nos. 2018YFC0806802 and 2018YFC0832105).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar]
  2. Berman, D.; Treibitz, T.; Avidan, S. Non-local image dehazing. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  3. Ren, W.; Ma, L.; Zhang, J.; Pan, J.; Cao, X.; Liu, W.; Yang, M.-H. Gated fusion network for single image dehazing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  4. Yang, D.; Sun, J. Proximal dehaze-net: A prior learning-based deep network for single image dehazing. In Proceedings of the The European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  5. Zhang, H.; Patel, V.M. Densely connected pyramid dehazing network. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  6. McCartney, E.J. Optics of the Atmosphere: Scattering by Molecules and Particles; John Wiley and Sons, Inc.: New York, NY, USA, 1976; p. 421. [Google Scholar]
  7. Narasimhan, S.G.; Nayar, S.K. Chromatic framework for vision in bad Weather. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2000 (Cat. No. PR00662), Hilton Head Island, SC, USA, 15 June 2000; pp. 598–605. [Google Scholar]
  8. Tan, R.T. Visibility in bad weather from a single image. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  9. Tarel, J.-P.; Hautiere, N. Fast visibility restoration from a single color or gray level image. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2201–2208. [Google Scholar]
  10. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel Prior. IEEE Trans. Pattern Anal. Mach. 2011, 33, 2341–2353. [Google Scholar]
  11. Meng, G.; Wang, Y.; Duan, J.; Xiang, S.; Pan, C. Efficient image dehazing with boundary constraint and contextual regularization. In Proceedings of the The IEEE International Conference on Computer Vision (ICCV), Sydney, Australia, 1–8 December 2013. [Google Scholar]
  12. Fattal, R. Dehazing using color-lines. ACM Trans. Graph. TOG 2014, 34, 13. [Google Scholar] [CrossRef]
  13. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An end-to-end system for single image haze removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Li, R.; Pan, J.; Li, Z.; Tang, J. Single image dehazing via conditional generative adversarial network. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  15. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.-H. Single image dehazing via multi-scale convolutional neural networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  16. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. Aod-net: All-in-one dehazing Network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778. [Google Scholar]
  17. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  18. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  19. Ma, K.; Liu, W.; Wang, Z. Perceptual evaluation of single image dehazing algorithms. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 3600–3604. [Google Scholar]
  20. Ancuti, C.; Ancuti, C.O.; Timofte, R.; de Vleeschouwer, C. I-haze: A dehazing benchmark with real hazy and haze-free indoor images. In International Conference on Advanced Concepts for Intelligent Vision Systems; Springer: Berlin/Heidelberg, Germany, 2018; pp. 620–631. [Google Scholar]
  21. Sakaridis, C.; Dai, D.; van Gool, L. Semantic foggy scene understanding with synthetic data. Int. J. Comput. Vis. 2018, 126, 973–992. [Google Scholar] [CrossRef] [Green Version]
  22. Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking single-image dehazing and beyond. IEEE Trans. Image Process. 2019, 28, 492–505. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  24. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
  25. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  26. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  27. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Smolley, S.P. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  28. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. arxiv 2016, arXiv:1611.07004. [Google Scholar]
  29. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  30. Zhang, H.; Xu, T.; Li, H.; Zhang, S.; Wang, X.; Huang, X.; Metaxas, D.N. Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  31. Yu, J.; Lin, Z.; Yang, J.; Shen, X.; Lu, X.; Huang, T.S. Generative image inpainting with contextual attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  32. Yang, X.; Xu, Z.; Luo, J. Towards perceptual image dehazing by physics-based disentanglement and adversarial training. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  33. Tarel, J.-P.; Hautiere, N.; Caraffa, L.; Cord, A.; Halmaoui, H.; Gruyer, D. Vision enhancement in homogeneous and heterogeneous fog. IEEE Intell. Transp. Syst. Mag. 2012, 4, 6–20. [Google Scholar] [CrossRef] [Green Version]
  34. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 746–760. [Google Scholar]
  35. Scharstein, D.; Szeliski, R. High-accuracy stereo depth maps using structured light. In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Madison, WI, USA, 18–20 June 2003; p. 1. [Google Scholar]
  36. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Instance normalization: The missing ingredient for fast stylization. arXiv 2016, arXiv:1607.08022. [Google Scholar]
  37. Xu, B.; Wang, N.; Chen, T.; Li, M. Empirical evaluation of rectified activations in convolutional network. arXiv 2015, arXiv:1505.00853. [Google Scholar]
  38. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  39. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  41. Noh, H.; Hong, S.; Han, B. Learning deconvolution network for semantic Segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1520–1528. [Google Scholar]
  42. Shi, W.; Caballero, J.; Huszar, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2016. [Google Scholar]
  43. Odena, A.; Dumoulin, V.; Olah, C. Deconvolution and checkerboard Artifacts. Distill 2016, 1, e3. [Google Scholar] [CrossRef]
  44. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  45. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 694–711. [Google Scholar]
  46. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  47. Ruderman, D.L. The statistics of natural images. Netw. Comput. Neural Syst. 2009, 5, 517–548. [Google Scholar] [CrossRef]
  48. Choi, L.K.; You, J.; Bovik, A.C. Referenceless prediction of perceptual fog density and perceptual image defogging. IEEE Trans. Image Process. 2015, 24, 3888–3901. [Google Scholar] [CrossRef]
  49. Ponomarenko, N.; Ieremeiev, O.; Lukin, V.; Egiazarian, K.; Jin, L.; Astola, J.; Vozel, B.; Chehdi, K.; Carli, M.; Battisti, F. Color image database tid2013: Peculiarities and preliminary results. In Proceedings of the European Workshop on Visual Information Processing, Paris, France, 10–12 June 2013. [Google Scholar]
  50. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  51. Huynh-Thu, Q.; Ghanbari, M. Scope of validity of psnr in image/video quality assessment. Electron. Lett. 2008, 4, 800–801. [Google Scholar] [CrossRef]
  52. Luo, M.R.; Cui, G.; Rigg, B. The development of the cie 2000 colour-difference formula: Ciede2000. Color Res. Appl. 2001, 26, 340–350. [Google Scholar] [CrossRef]
  53. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  54. Liu, L.; Liu, B.; Huang, H.; Bovik, A.C. No-reference image quality assessment based on spatial and spectral entropies. Sig. Process Image Commun. 2014, 29, 856–863. [Google Scholar] [CrossRef]
  55. Saad, M.A.; Bovik, A.C.; Charrier, C. Blind image quality assessment: A natural scene statistics approach in the DCT domain. IEEE Trans. Image Process. 2012, 21, 3339–3352. [Google Scholar] [CrossRef]
  56. Esfandarani, H.T.; Milanfar, P. NIMA: Neural image assessment. IEEE Trans. Image Process. 2018, 27, 3998–4011. [Google Scholar]
Figure 1. Overwater image dehazing example. The proposed method generates more clear images compared to state-of-the-art methods. (a) Hazy input, (b) Cai [13], (c) Yang [4], (d) ours.
Figure 1. Overwater image dehazing example. The proposed method generates more clear images compared to state-of-the-art methods. (a) Hazy input, (b) Cai [13], (c) Yang [4], (d) ours.
Electronics 09 01877 g001
Figure 2. The main architecture of the proposed OverWater Image (OWI)-Dehazing network. G and F denote generators, where G : X Y generates clean images from hazy images and F : Y X vice versa. D x and D y denote discriminators. Adversarial loss, cycle consistency loss and perceptual loss are employed to train the network.
Figure 2. The main architecture of the proposed OverWater Image (OWI)-Dehazing network. G and F denote generators, where G : X Y generates clean images from hazy images and F : Y X vice versa. D x and D y denote discriminators. Adversarial loss, cycle consistency loss and perceptual loss are employed to train the network.
Electronics 09 01877 g002
Figure 3. Architecture of our generator and discriminator. The generator consists of encoding, transformation, and decoding three parts.
Figure 3. Architecture of our generator and discriminator. The generator consists of encoding, transformation, and decoding three parts.
Electronics 09 01877 g003
Figure 4. Examples of mean substracted contrast normalization (MSCN) coefficient.
Figure 4. Examples of mean substracted contrast normalization (MSCN) coefficient.
Electronics 09 01877 g004
Figure 5. Architecture of the proposed image quality assessment (IQA) model for dehazed images.
Figure 5. Architecture of the proposed image quality assessment (IQA) model for dehazed images.
Electronics 09 01877 g005
Figure 6. Examples of HazyWater dataset (best viewed in color). (af) Hazy images. (gl) Clean imags.
Figure 6. Examples of HazyWater dataset (best viewed in color). (af) Hazy images. (gl) Clean imags.
Electronics 09 01877 g006
Figure 7. Real light hazy images and corresponding dehazing results from several state-of-the-art methods (best viewed in color). (a) Hazy Image. (b) DCP. (c) FVR. (d) BCCR. (e) CAP. (f) DehazeNet. (g) MSCNN. (h) AOD-Net. (i) dehaze-cGAN (j) Proximal. (k) CycleGAN. (l) Ours.
Figure 7. Real light hazy images and corresponding dehazing results from several state-of-the-art methods (best viewed in color). (a) Hazy Image. (b) DCP. (c) FVR. (d) BCCR. (e) CAP. (f) DehazeNet. (g) MSCNN. (h) AOD-Net. (i) dehaze-cGAN (j) Proximal. (k) CycleGAN. (l) Ours.
Electronics 09 01877 g007
Figure 8. Real heavy hazy images and corresponding dehazing results from several state-of-the-art methods (best viewed in color). (a) Hazy Image. (b) Dark-channel prior (DCP). (c) FVR. (d) BCCR. (e) Color attenuation prior (CAP). (f) DehazeNet. (g) MSCNN. (h) AOD-Net. (i) dehaze-cGAN (j) Proximal. (k) CycleGAN. (l) Ours.
Figure 8. Real heavy hazy images and corresponding dehazing results from several state-of-the-art methods (best viewed in color). (a) Hazy Image. (b) Dark-channel prior (DCP). (c) FVR. (d) BCCR. (e) Color attenuation prior (CAP). (f) DehazeNet. (g) MSCNN. (h) AOD-Net. (i) dehaze-cGAN (j) Proximal. (k) CycleGAN. (l) Ours.
Electronics 09 01877 g008
Figure 9. Comparison in terms of SSIM/PSNR for different image dehazing methods. (a) Ground Truth. (b) DCP. (c) FVR. (d) BCCR. (e) CAP. (f) DehazeNet. (g) MSCNN. (h) AOD-Net. (i) dehaze-cGAN. (j) Proximal. (k) CycleGAN. (l) Ours.
Figure 9. Comparison in terms of SSIM/PSNR for different image dehazing methods. (a) Ground Truth. (b) DCP. (c) FVR. (d) BCCR. (e) CAP. (f) DehazeNet. (g) MSCNN. (h) AOD-Net. (i) dehaze-cGAN. (j) Proximal. (k) CycleGAN. (l) Ours.
Electronics 09 01877 g009
Figure 10. Examples of paired data generated by our network. (ad) Clean images. (eh) Generated hazy images corresponding to (ad).
Figure 10. Examples of paired data generated by our network. (ad) Clean images. (eh) Generated hazy images corresponding to (ad).
Electronics 09 01877 g010
Figure 11. Comparison via the proposed IQA model. Ranking scores and the rankings are shown below each image. (a) Ours. (b) CAP. (c) MSCNN. (d) DehazeNet. (e) Proximal. (f) DCP. (g) Input. (h) BCCR. (i) FVR. (j) AOD-Net.
Figure 11. Comparison via the proposed IQA model. Ranking scores and the rankings are shown below each image. (a) Ours. (b) CAP. (c) MSCNN. (d) DehazeNet. (e) Proximal. (f) DCP. (g) Input. (h) BCCR. (i) FVR. (j) AOD-Net.
Electronics 09 01877 g011
Figure 12. Examples of color cast. (a,f) Two hazy images. (b,g) Generated by DCP. (c,h) Generated by DehazeNet. (d,i) Generated by MSCNN. (e,j) Our results. Compared with DCP, DehazeNet and MSCNN, our method effectively corrects the color cast phenomenon and produces a visual better image.
Figure 12. Examples of color cast. (a,f) Two hazy images. (b,g) Generated by DCP. (c,h) Generated by DehazeNet. (d,i) Generated by MSCNN. (e,j) Our results. Compared with DCP, DehazeNet and MSCNN, our method effectively corrects the color cast phenomenon and produces a visual better image.
Electronics 09 01877 g012
Figure 13. Effectiveness of the proposed network with resize convolution. (a,e) Input hazy images. (b,f) Dehazed images using deconvolution. (c,g) The dehazed images using sub-pixel convolution. (d,h) Denote dehazed results using resize convolution (ours). (eh) The zoom-in views of (ad), separately.
Figure 13. Effectiveness of the proposed network with resize convolution. (a,e) Input hazy images. (b,f) Dehazed images using deconvolution. (c,g) The dehazed images using sub-pixel convolution. (d,h) Denote dehazed results using resize convolution (ours). (eh) The zoom-in views of (ad), separately.
Electronics 09 01877 g013
Figure 14. Comparison of dehazing with and without perceptual loss. (ae) The generation direction of X Y X . (fj) The direction of formation is Y X Y . “(w/o L p e r )” denotes the network without perceptual loss, and “(w/ L p e r )” denotes the network with perceptual loss.
Figure 14. Comparison of dehazing with and without perceptual loss. (ae) The generation direction of X Y X . (fj) The direction of formation is Y X Y . “(w/o L p e r )” denotes the network without perceptual loss, and “(w/ L p e r )” denotes the network with perceptual loss.
Electronics 09 01877 g014
Figure 15. Example of failure cases. (a) Input image. (b) CAP. (c) DehazeNet. (d) MSCNN. (e) Ours.
Figure 15. Example of failure cases. (a) Input image. (b) CAP. (c) DehazeNet. (d) MSCNN. (e) Ours.
Electronics 09 01877 g015
Table 1. Main parameters of our generator. ‘#In’ denotes the number of input channel, and ‘#Out’ represents the number of output channel. ‘Upsample’ denotes the resize convolution.
Table 1. Main parameters of our generator. ‘#In’ denotes the number of input channel, and ‘#Out’ represents the number of output channel. ‘Upsample’ denotes the resize convolution.
Layer Name#In#OutKernelStrides
Conv 1332 7 × 7 1
Conv 23264 3 × 3 2
Conv 364128 3 × 3 2
Res Block 1128138 3 × 3 1
...............
Res Block 9128128 3 × 3 1
Upsample 125664 3 × 3 2
Upsample 212832 3 × 3 2
Conv 4323 7 × 7 1
Tanh----
Table 2. Main parameters of our discriminator. ‘#In’ denotes the number of input channel, and ‘#Out’ represents the number of output channel.
Table 2. Main parameters of our discriminator. ‘#In’ denotes the number of input channel, and ‘#Out’ represents the number of output channel.
Layer Name#In#OutKernelStrides
Conv 1332 4 × 4 2
Conv 23264 4 × 4 2
Conv 364128 4 × 4 2
Conv 4128256 4 × 4 2
Conv 52561 4 × 4 1
Table 3. Average PSNR, SSIM, and CIEDE2000 values of dehazed results on the new synthetic dataset. The best result, the second result, and the third place result are represented by red, blue, and green, respectively. The average values of PSNR, SSIM and CIEDE2000 results calculated directly between the each hazy and its corresponding clean image.
Table 3. Average PSNR, SSIM, and CIEDE2000 values of dehazed results on the new synthetic dataset. The best result, the second result, and the third place result are represented by red, blue, and green, respectively. The average values of PSNR, SSIM and CIEDE2000 results calculated directly between the each hazy and its corresponding clean image.
FVRDCPBCCRCAPDehazeNetMSCNNAOD-NetProximaldehaze-cGANCycleGANOurs
SSIM0.8170.7170.6800.8250.8670.8500.8350.8200.8760.5840.893
PSNR16.5614.5713.9219.5023.2720.5519.3317.7922.7718.3124.93
CIEDE200011.7114.9215.359.376.237.819.1110.146.5111.55.98
Table 4. Comparison of dehazed image quality using four image quality assessment methods. The top three results are in red, blue, and green font, respectively.
Table 4. Comparison of dehazed image quality using four image quality assessment methods. The top three results are in red, blue, and green font, respectively.
FADE [48] ↓SSEQ [54] ↓BLINDS-2 [55] ↓NIMA [56] ↑
Ours1.9536.2431.506.47
CAP3.0340.9751.005.95
MSCNN2.3845.5650.506.23
DehazeNet3.1142.4949.506.12
Proximal-Dehaze1.6943.6954.506.06
DCP1.4543.3251.505.36
BCCR1.1745.7648.506.08
FVR2.6741.7148.004.86
AOD-Net2.3038.6254.506.15
Table 5. Average run time (in seconds) on synthetic data. The top three results are highlighted in red, blue, and green font, respectively.
Table 5. Average run time (in seconds) on synthetic data. The top three results are highlighted in red, blue, and green font, respectively.
MethodsTime ( 512 × 512 )Run Environment
DCP15.55Matlab
FVR3.87Matlab
CAP0.94Matlab
DehazeNet1.02Matlab
MSCNN1.97Matlab
AOD-Net0.23PyTorch
dehaze-cGAN0.64Torch7
Ours0.30TensorFlow
Table 6. Average scores in terms of PSNR, SSIM, and CIEDE2000 for deconvolution, sub-pixel convolution, and resize convolution on the synthetic test set from HazyWater Dataset.
Table 6. Average scores in terms of PSNR, SSIM, and CIEDE2000 for deconvolution, sub-pixel convolution, and resize convolution on the synthetic test set from HazyWater Dataset.
Average MetricsSSIMPSNRCIEDE2000
Deconvolution0.75820.349.12
Sub-pixel Convolution0.64320.2110.86
Resize Convolution0.81922.197.41
Table 7. Effect of perceptual loss in terms of SSIM, PSNR, and CIEDE2000. CycleGAN loss refers to the formulation of adversarial loss and cycle consistency loss, VGG loss refers to substitute perceptual loss.
Table 7. Effect of perceptual loss in terms of SSIM, PSNR, and CIEDE2000. CycleGAN loss refers to the formulation of adversarial loss and cycle consistency loss, VGG loss refers to substitute perceptual loss.
Average MetricsSSIMPSNRCIEDE2000
CycleGAN loss0.81922.197.41
CycleGAN loss + VGG loss0.89324.935.98
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zheng, S.; Sun, J.; Liu, Q.; Qi, Y.; Yan, J. Overwater Image Dehazing via Cycle-Consistent Generative Adversarial Network. Electronics 2020, 9, 1877. https://doi.org/10.3390/electronics9111877

AMA Style

Zheng S, Sun J, Liu Q, Qi Y, Yan J. Overwater Image Dehazing via Cycle-Consistent Generative Adversarial Network. Electronics. 2020; 9(11):1877. https://doi.org/10.3390/electronics9111877

Chicago/Turabian Style

Zheng, Shunyuan, Jiamin Sun, Qinglin Liu, Yuankai Qi, and Jianen Yan. 2020. "Overwater Image Dehazing via Cycle-Consistent Generative Adversarial Network" Electronics 9, no. 11: 1877. https://doi.org/10.3390/electronics9111877

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop