Next Article in Journal
Powering the Environmental Internet of Things
Next Article in Special Issue
Violence Detection Using Spatiotemporal Features with 3D Convolutional Neural Network
Previous Article in Journal
Anti-Multipath Performance Improvement of an M-ary Position Phase Shift Keying Modulation System
Previous Article in Special Issue
Spatio–Temporal Image Representation of 3D Skeletal Movements for View-Invariant Action Recognition with Deep Convolutional Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SCENet: Secondary Domain Intercorrelation Enhanced Network for Alleviating Compressed Poisson Noises

IoT Research Division, Hyper-connected Communication Research Lab., Electronics Telecommunications Research Institute (ETRI), Daejeon 34129, Korea
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(8), 1939; https://doi.org/10.3390/s19081939
Submission received: 23 March 2019 / Revised: 17 April 2019 / Accepted: 24 April 2019 / Published: 25 April 2019
(This article belongs to the Special Issue Deep Learning-Based Image Sensors)

Abstract

:
In real image coding systems, block-based coding is often applied on images contaminated by camera sensor noises such as Poisson noises, which cause complicated types of noises called compressed Poisson noises. Although many restoration methods have recently been proposed for compressed images, they do not provide satisfactory performance on the challenging compressed Poisson noises. This is mainly due to (i) inaccurate modeling regarding the image degradation, (ii) the signal-dependent noise property, and (iii) the lack of analysis on intercorrelation distortion. In this paper, we focused on the challenging issues in practical image coding systems and propose a compressed Poisson noise reduction scheme based on a secondary domain intercorrelation enhanced network. Specifically, we introduced a compressed Poisson noise corruption model and combined the secondary domain intercorrelation prior with a deep neural network especially designed for signal-dependent compression noise reduction. Experimental results showed that the proposed network is superior to the existing state-of-the-art restoration alternatives on classical images, the LIVE1 dataset, and the SIDD dataset.

1. Introduction with Preliminary Examination

The block-based discrete cosine transform (BDCT) coding scheme is typically adopted for various coding standards such as JPEG, MPEG4, H.264/AVC, and H.265/HEVC for image and video compression. However, block-based coding suffers from well-known undesirable blocking artifacts due to the distortion of spatial correlation between neighboring blocks called intercorrelation. Meanwhile, according to [1], noises in real images captured by charge-coupled device (CCD) imaging sensors generally tend to have signal-dependent characteristics such as a Poisson distribution. Accordingly, the coding of the Poisson noise-corrupted image generates complex signal-dependent compressed noises that are called compressed Poisson noises. By applying recent compressed image restoration algorithms [2,3,4,5,6,7,8,9,10,11,12,13,14] based on convolutional neural networks (CNN) on the degraded image, it may be possible to reduce the conventional blocking and ringing artifacts. However, despite their promising solutions on conventional artifacts, the existing algorithms still do not provide excellent performance on compressed Poisson noises. This is because both an accurate image degradation model considering practical imaging systems and signal-dependent noise characteristics have not been seriously dealt with in the existing neural networks. To cope with the issues due to compressed Poisson noises, in this paper, we introduced an image degradation model suitable to the practical application and present a robust multi-band neural network to the signal-dependent noise property by exploiting a variance-stabilizing transformation (VST) [15]. In addition, the existing restoration algorithms [2,3,4,5,6,7,8,9,10,11,12,13,14] may not assure their best performance on coded images especially at low bit rates because the intercorrelation distortion recovery of each BDCT coefficient is not analytically reflected in the existing networks. In this paper, to verify the effect of the block-based coding on the intercorrelation, we performed a preliminary study by investigating the intercorrelation of each BDCT coefficient between neighboring blocks. To this end, we first prepared four ground truth (GT) images given in the top left of Figure 1 and produced their JPEG-coded images with quality factors, q, of 10, 20, and 30 corresponding to low bit rates. For the four GT images and their twelve coded images, we obtain the secondary block, b, which is composed of 4 × 4 BDCT coefficients [16]. We then computed the coefficient’s spatial correlation in each secondary block, or the intercorrelation ρ as
ρ = min ( | R 0 , 1 R 0 , 0 | ,   | R 1 , 0 R 0 , 0 | )
where
R x , y = m = 0 3 x n = 0 3 y b ( m , n ) b ( m + x , n + y ) ( 4 x ) ( 4 y ) .
Note that we investigated the intercorrelation using both GT images and the corresponding coded images, while the previous study [16] only used GT images. We further observed the change of the intercorrelation distribution according to the compression level of the coded images, unlike in an earlier study [17].
Figure 1 shows the distributions of the computed high intercorrelation values (ρ > 0.6) for the three lowest frequency (LF) BDCT coefficients: DC, AC(1,0), and AC(0,1). As the compression level increases, the relative frequencies of the high intercorrelation values for three LF coefficients commonly decreases. For example, the relative frequency of 63.4% for the DC of GT images decreases to 58.0%, 52.5%, and 40.5% in JPEG-coded images with q of 30, 20, and 10, respectively. The same tendency could also be observed for the other two LF coefficients, AC(1,0) and AC(0,1), while the remaining high-frequency (HF) coefficients did not necessarily follow this trend. From this examination, we considered that the intercorrelation enhancement of three LF coefficients was required for the effective restoration of coded images. In addition, the intercorrelation distortion of the BDCT coefficients occurred differently for each coefficient, as a different quantization step size was applied for each one in JPEG. Hence, in this paper, we propose an intercorrelation enhancement network in the secondary domain, which enabled us to improve the distorted intercorrelation of each frequency coefficient adaptively.

2. Degradation Model

Considering the practical imaging systems described above, we defined a simple but effective image degradation model consisting of three procedures: camera sensor noise corruption, image coding based on quantization in the BDCT domain, and image decoding based on dequantization in the BDCT domain, as illustrated in Figure 2.
Here, let z be a decoded block in the receiver, specifically,
z ( x ) = T 1 { r o u n d { T { a P ( x ) } Q } × Q }
where x denotes a spatial coordinate; and T and T−1 are the BDCT and inverse BDCT (IBDCT) operators, respectively. In addition, P denotes a Poisson variable scaled by a with a mean value µ, and Q denotes a quantization table. The probability distribution of the acquired value P(x) is derived as
p ( P ( x ) | μ ) = μ P ( x ) e μ P ( x ) ! .
To better observe the visual image degradation of the received image z, a residual image can be obtained by subtracting the GT image from z, as shown in the top right of Figure 2. We note here that the decoded image suffers from complicated degradations including compressed Poisson noises as well as well-known blocking and ringing artifacts. It can also be observed that original near-random Poisson noises in P(x) were deformed to annoying patterns that had a strong spatial correlation, even in smooth regions.

3. Secondary Domain Intercorrelation Enhanced Network

To reflect the defined image degradation model in the neural network adequately, in this paper, we suggested a secondary domain intercorrelation enhanced network (SCENet), which is quite suitable to address compressed Poisson noises, as illustrated in Figure 3. Inspired by our preliminary examination, we adapted and extended the secondary domain approach [17], which is still valuable for recovering the intercorrelation distortion. Note that the proposed algorithm has a clear difference from the existing algorithm [17] in terms of method and application. Specifically, while the classical edge-preserving total variation (TV) filtering was applied in the secondary domain for removing well-known blocking artifacts in [17], we utilized the deep neural network specially trained for reducing compressed Poisson noises, instead of the classical filtering. We also exploited the variance-stabilizing model to deal with signal-dependent noise characteristics, unlike in [17]. In other words, we combined key elements of the VST-based secondary domain intercorrelation model with a deep neural network that was particularly trained using the defined compressed Poisson noise model. The proposed SCENet architecture had two major parallel phases: restoration of the three LF coefficients and restoration of the high-band (HB) image.
In one of the parallel phases, we increased the intercorrelation of DC, AC(1,0), and AC(0,1) in the secondary domain, as shown at the top of Figure 3. In particular, the network had 20 layer architectures and each layer architecture was composed of five operations: VST, convolution, inverse VST (IVST), batch normalization (BN), and rectified linear units (ReLU). For an input image z, we first generated three secondary images: SDC, SAC(1,0), and SAC(0,1). To this end, we computed three LF BDCT coefficients in each 8 × 8 block by shifting the block pixel-by-pixel with overlapping and then merged them into each secondary image, respectively. After that, to remove the signal dependency of compressed noises, the secondary image pixel value s was stabilized to have homoscedastic variance via the Anscombe transformation [15]
f ( s ) = { 2 s + 3 / 8   , s 3 / 8 0    , otherwise .
As a subsequent procedure, the convolution was undertaken with K pre-trained filters, aL,F, with a size of W × H. Next, destabilization based on the IVST was applied in order to retrieve the original heteroskedastic variance as
E [ f ( s ) | μ ] = i = 0 ( 2 s + 3 / 8 μ s e μ s ! ) .
The IVST step was then followed by BN and ReLU for fast and stable convergence in the training process. The iterative layer architectures were performed on three secondary images separately for the adaptive restoration of each coefficient that had different quantization amounts. Final feature maps were reshaped to the original input tensor size via a fully-connected layer and then an output low-band (LB) image Lout was obtained by applying T−1 to three filtered coefficients, SDC,out, SAC(1,0),out, and SAC(0,1),out, in each block without overlapping. The images from the first column to the fourth column of Figure 4 show that the three LF coefficients restoration network successfully recovered the secondary images and the LB image, similar to their corresponding GT images by addressing artifacts in degraded images.
In another parallel phase for restoring the HB image, we first obtained the input LB image L by applying T−1 to three LF coefficients in each 8 × 8 block. The input HB image H was then acquired by subtracting L from the input image z. Note that H corresponded to the remaining 61 HF coefficients in each block. Next, the filtered output Hout was computed via the same iterative layer architectures as in the restoration of the three LF coefficients by using convolution filters bL instead of aL,F. Note in the fifth column of Figure 4 that the network effectively recovered Hout quite close to the GT. The final restoration result could be obtained by adding two output images, Lout and Hout. All of the above steps are also described in Algorithm 1.
Algorithm 1: Compressed Poisson noise reduction based on the SCENet
Input: degraded image z and trained parameters aL=1,...,20, F=1,2,3, bL=1,...,20
Output: restored image y
1:  Compute DC, AC(1,0), AC(0,1) by T{z}.
2:  Obtain S = {SDC, SAC(1,0), SAC(0,1)} by merging the each coefficient.
3:  for L = 1, …, 20 do
4:  Stabilize using VST by f{S}.
5:  Apply convolution with trained parameters and then destabilize it by f −1{aL,F* f{S}}.
6:  Appy BN and ReLU by max(BN{f −1{aL,F* f{S}}}, 0).
7:  end for
8:  Obtain Sout = {SDC,out, SAC(1,0),out, SAC(0,1),out} by applying a fully-connected layer.
9:  Estimate Lout by T−1{Sout}.
10: Obtain H by zT−1{S}.
11: Estimate Hout by running steps 3−8 above with H and bL=1,...,20 instead of S and aL,F.
12: Estimate final restored image by y = Lout + Hout.

4. Experiments

To train our networks, we used 400 images from the BSDS500 dataset [18] and 800 images from the DIV2K dataset [19]. Given a GT image, we synthesized three JPEG degraded images with different noise levels of {quality factor q, peak} = {10, 200}, {20, 400}, and {30, 600}. To generate Poisson noises, the maximum intensity of the GT image was first normalized to have the defined peak value. Next, the noise corruption was performed on the normalized image and the corrupted image was then denormalized to have the original maximum intensity. Therefore, the lower the peak value, the higher the Poisson noise level. In the restoration network of DC, AC(1,0), and AC(0,1), the sizes of W × H for filter parameters aL,F = 1, aL,F = 2, and aL,F = 3 were set to 3 × 3, 3 × 1, and 1 × 3, respectively, by considering the dominant pattern of compression artifacts in each secondary image. For example, the secondary image of AC(1,0) (or AC(0,1)) included only the vertical (or horizontal) artifacts affected by the BDCT basis function, as shown in the second row of Figure 4. In addition, the number of filters K in all networks was set to 64 and the W × H for the HB image restoration network parameters bL was empirically set to 3 × 3 in every architecture layer. Given a set of secondary images and HB images computed from GT images and their corresponding degraded images, we used the mean squared error (MSE) as a loss function. To minimize the loss function, we adopted an optimization method, Adam [20] with a batch size of 32. The learning rate was set to drop exponentially from 1e−3 to 1e−5. The proposed network was separately trained according to each noise level on one NVIDIA GTX 1080 GPU, under MATLAB R2017b with the MatConvNet package for about 16 h. Figure 5 shows an example of the convolution filters that were obtained via the network training. The whole inference time was about 120 ms for a 512 × 512 image and the time could be further reduced via parallel processing.
Meanwhile, in order to validate our trained networks, we used eight classical images given in Table 1 and 29 images from the LIVE1 dataset [21]. Figure 6 and Figure 7 show several restoration results for the JPEG degraded images with different noise levels of q and peak values to evaluate the subjective performance of the proposed network. We also compared the performance with a general compression artifact reduction algorithm [2] and two existing state-of-the-art restoration algorithms [4,5] based on CNN. The two existing denoising algorithms [22,23] that were not based on CNN were additionally used for the comparison. Open source codes in the first authors’ websites were applied for the comparison. The pre-trained models for MWCNN [5] were kindly provided by P. Liu, because it was not accessible via that website. We can easily note that in Figure 6 and Figure 7 that the results of the existing algorithms were not satisfactory because the undesirable compressed Poisson noises still remained, especially in many flat regions such as the wing, the face, the pepper, the calendar, the sky, and the wall. In contrast, the proposed SCENet provided more visually pleasing images by successfully alleviating the annoying compressed Poisson noises while preserving the image details in comparison to the existing algorithms. This noticeable visual improvement was achieved by the proposed VST-based secondary domain intercorrelation prior that was enforced in the neural network. The full resolution image results and an executable program for reproducing the results are also available on our website [24].
In addition to the subjective comparison, a quantitative comparison was conducted. Table 1 summarizes the peak signal-to-noise ratio (PSNR) and structure similarity (SSIM) [25] values computed from the processed results of the eight classical test images. It can be noticed in the table that the proposed SCENet provided the best objective quality except for only one case, by restoring the GT pixel values well. The objective comparison on the LIVE1 dataset was additionally conducted, as given in Table 2. The average PSNR and SSIM values were calculated from the luminance channels of 29 images in the dataset. This demonstrates that the proposed network overall outperformed the existing compressed image restoration algorithms as well as providing significant quality improvement when compared with the input degraded images.
In addition, to evaluate the restoration performance of several algorithms on the actual sensor noises, we used the smartphone image denoising dataset (SIDD) [26] because smartphone images tend to have notably severe Poisson noises due to the small aperture and sensor size. Figure 8 and Figure 9 show the comparison of the algorithms for two images from the SIDD dataset, Books and Desk, respectively. The two images were acquired using an iPhone 7 with different camera settings, ISO, and exposure time and they included real camera sensor noises, as shown in Figure 8a and Figure 9a. The JPEG compression of the sensor noises generates compressed Poisson noises, as shown in Figure 8b and Figure 9b. We note that the proposed SCENet alleviated the compressed Poisson noises well, as shown in Figure 8h and Figure 9h, while the results of the existing algorithms still included the noises in the book, the phone, the paper, and the box, as shown in Figure 8c–g and Figure 9c–g. As their original GT images were not available, we conducted an objective comparison using a no-reference image quality metric, the blind/referenceless image spatial quality evaluator (BRISQUE) index [27]. The lower the BRISQUE values, the better the image quality. As expected in visual results in Figure 8 and Figure 9, we noticed that SCENet outperformed the existing algorithms on the real noise data in terms of the BRISQUE values.

5. Conclusions

Compressed Poisson noises are critical and troublesome issues generated in real image coding systems. To address sensor noises effectively, we analyzed the intercorrelation distortion process via our preliminary examination and proposed a new multi-band intercorrelation increment network that exploits the secondary domain instead of the typical spatial domain. Additionally, to increase robustness to the signal-dependent noise characteristics, we designed a layer architecture composed of five operations and trained the network parameters under the challenging image degradation model. The superior performance of the proposed network on three datasets was also validated in terms of both subjective and objective qualities.

Author Contributions

Conceptualization, S.B.Y. and M.H.; Methodology, S.B.Y.; Software, S.B.Y.; Validation, S.B.Y.; Formal Analysis, S.B.Y.; Investigation, S.B.Y.; Resources, S.B.Y.; Data Curation, S.B.Y.; Writing—Original Draft Preparation, S.B.Y.; Writing—Review & Editing, S.B.Y.; Visualization, S.B.Y.; Supervision, M.H.; Project Administration, M.H.; Funding Acquisition, M.H.

Funding

This work was supported by ‘The Cross-Ministry Giga KOREA Project’ grant funded by the Korean government (MSIT) (No. GK18P0600, Development and Demonstration of Smart City Service over 5G Network).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Foi, A.; Trimeche, M.; Katkovnik, V.; Egiazarian, K. Practical Poissonian-Gaussian noise modeling and fitting for single-image raw-data. IEEE Trans. Image Process. 2008, 17, 1737–1754. [Google Scholar] [CrossRef] [PubMed]
  2. Dong, C.; Deng, Y.; Change, L.C.; Tang, X. Compression artifacts reduction by a deep convolutional network. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 576–584. [Google Scholar]
  3. Chen, Y.; Pock, T. Trainable nonlinear reaction diffusion: A flexible framework for fast and effective image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1256–1272. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, P.; Zhang, H.; Zhang, K.; Lin, L.; Zuo, W. Multi-level wavelet-CNN for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 773–782. [Google Scholar]
  6. Lu, G.; Ouyang, W.; Xu, D.; Zhang, X.; Gao, Z.; Sun, M.T. Deep Kalman filtering network for video compression artifact reduction. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 568–584. [Google Scholar]
  7. Zhang, Y.; Sun, L.; Yan, C.; Ji, X.; Dai, Q. Adaptive residual networks for high-quality image restoration. IEEE Trans. Image Process. 2018, 27, 3150–3163. [Google Scholar] [CrossRef] [PubMed]
  8. Zhang, X.; Yang, W.; Hu, Y.; Liu, J. DMCNN: Dual-domain multi-scale convolutional neural network for compression artifacts removal. In Proceedings of the 25th IEEE International Conference on Image Processing, Athens, Greece, 7–10 October 2018; pp. 390–394. [Google Scholar]
  9. Zheng, B.; Chen, Y.; Tian, X.; Zhou, F.; Liu, X. Implicit dual-domain convolutional network for robust color image compression artifact reduction. arXiv 2018, arXiv:1810.08042. [Google Scholar]
  10. Zhang, X.; Lu, Y.; Liu, J.; Dong, B. Dynamically unfolding recurrent restorer: A moving endpoint control method for image restoration. arXiv 2018, arXiv:1805.07709. [Google Scholar]
  11. Zheng, B.; Sun, R.; Tian, X.; Chen, Y. S-Net: A scalable convolutional neural network for JPEG compression artifact reduction. J. Electron. Imaging 2018, 27, Art. No. 043037. [Google Scholar]
  12. Yoo, J.; Lee, S.H.; Kwak, N. Image restoration by estimating frequency distribution of local patches. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6684–6692. [Google Scholar]
  13. Galteri, L.; Seidenari, L.; Bertini, M.; Bimbo, A.D. Deep universal generative adversarial compression artifact removal. IEEE Trans. Multimedia 2019, in press. [Google Scholar] [CrossRef]
  14. Chen, H.; He, X.; An, C.; Nguyen, T.Q. Deep wide-activated residual network based joint blocking and color bleeding artifacts reduction for 4: 2: 0 JPEG-compressed images. IEEE Signal Process. Lett. 2019, 26, 79–83. [Google Scholar] [CrossRef]
  15. Markku, M.; Foi, A. Optimal inversion of the Anscombe transformation in low-count Poisson image denoising. IEEE Trans. Image Process. 2011, 20, 99–109. [Google Scholar]
  16. Lim, K.W.; Chun, K.W.; Ra, J.B. Improvements on image transform coding by reducing interblock correlation. IEEE Trans. Image Process. 1995, 4, 1146–1150. [Google Scholar] [PubMed]
  17. Yoo, S.B.; Choi, K.; Ra, J.B. Post-processing for blocking artifact reduction based on inter-block correlation. IEEE Trans. Multimedia 2014, 16, 1536–1548. [Google Scholar] [CrossRef]
  18. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; pp. 416–423. [Google Scholar]
  19. Agustsson, E.; Timofte, R. Ntire 2017 challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 126–135. [Google Scholar]
  20. Kingma, D.P.; Adam, B.J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  21. LIVE Image Quality Assessment Database Release 2. Available online: http://live.ece.utexas.edu/research/quality (accessed on 15 March 2019).
  22. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  23. Zhao, C.; Zhang, J.; Ma, S.; Fan, X.; Zhang, Y.; Gao, W. Reducing image compression artifacts by structural sparse representation and quantization constraint prior. IEEE Trans. Circuits Sys. Video Technol. 2017, 27, 2057–2071. [Google Scholar] [CrossRef]
  24. SCENet Program. Available online: https://github.com/seokbongyoo/SCENet (accessed on 15 March 2019).
  25. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  26. Smartphone Image Denoising Dataset. Available online: https://www.eecs.yorku.ca/~kamel/sidd/index.php (accessed on 5 April 2019).
  27. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Intercorrelation distribution according to the compression level.
Figure 1. Intercorrelation distribution according to the compression level.
Sensors 19 01939 g001
Figure 2. Image degradation model due to the coding of Poisson noisy images.
Figure 2. Image degradation model due to the coding of Poisson noisy images.
Sensors 19 01939 g002
Figure 3. The architecture of the secondary domain intercorrelation enhanced network.
Figure 3. The architecture of the secondary domain intercorrelation enhanced network.
Sensors 19 01939 g003
Figure 4. Stepwise restoration results of the SCENet. The first, second, and third columns show secondary images of DC, AC(1,0), and AC(0,1), respectively. The fourth and fifth columns show the LB and HB images, respectively.
Figure 4. Stepwise restoration results of the SCENet. The first, second, and third columns show secondary images of DC, AC(1,0), and AC(0,1), respectively. The fourth and fifth columns show the LB and HB images, respectively.
Sensors 19 01939 g004
Figure 5. An example of convolution filter parameters that were obtained via our network training.
Figure 5. An example of convolution filter parameters that were obtained via our network training.
Sensors 19 01939 g005
Figure 6. Comparison of the subjective visual quality for four classical images: Butterfly, Lena, Peppers, and Mobile. (a) Ground truths. (b) Degraded images with q = 10 and peak = 200. (c). Results of BM3D [22]. (d) Results of SSRQC [23]. (e) Results of ARCNN [2]. (f) Results of DnCNN [4]. (g) Results of MWCNN [5]. (h) Results of SCENet.
Figure 6. Comparison of the subjective visual quality for four classical images: Butterfly, Lena, Peppers, and Mobile. (a) Ground truths. (b) Degraded images with q = 10 and peak = 200. (c). Results of BM3D [22]. (d) Results of SSRQC [23]. (e) Results of ARCNN [2]. (f) Results of DnCNN [4]. (g) Results of MWCNN [5]. (h) Results of SCENet.
Sensors 19 01939 g006
Figure 7. Comparison of the subjective visual quality for four images from the LIVE1 database: Caps, Womanhat, Carnivaldolls, and Lighthouse2. (a) Ground truths. (b) Degraded images with q = 20 and peak = 400. (c) Results of BM3D [22]. (d) Results of SSRQC [23]. (e) Results of ARCNN [2]. (f) Results of DnCNN [4]. (g) Results of MWCNN [5]. (h) Results of SCENet.
Figure 7. Comparison of the subjective visual quality for four images from the LIVE1 database: Caps, Womanhat, Carnivaldolls, and Lighthouse2. (a) Ground truths. (b) Degraded images with q = 20 and peak = 400. (c) Results of BM3D [22]. (d) Results of SSRQC [23]. (e) Results of ARCNN [2]. (f) Results of DnCNN [4]. (g) Results of MWCNN [5]. (h) Results of SCENet.
Sensors 19 01939 g007
Figure 8. Comparison of the subjective visual quality for the image from the SIDD dataset, Books. (a) Captured image using an iPhone 7 with ISO = 800 and exposure = 1/2000 (BRISQUE = 43.36). (b) JPEG-coded image with q = 30 (BRISQUE = 41.97). (c) Result of BM3D [22] (BRISQUE = 45.06). (d) Result of SSRQC [23] (BRISQUE = 37.01). (e) Result of ARCNN [2] (BRISQUE = 33.15). (f) Result of DnCNN [4] (BRISQUE = 33.09). (g) Result of MWCNN [5] (BRISQUE = 37.65). (h) Result of SCENet (BRISQUE = 25.55).
Figure 8. Comparison of the subjective visual quality for the image from the SIDD dataset, Books. (a) Captured image using an iPhone 7 with ISO = 800 and exposure = 1/2000 (BRISQUE = 43.36). (b) JPEG-coded image with q = 30 (BRISQUE = 41.97). (c) Result of BM3D [22] (BRISQUE = 45.06). (d) Result of SSRQC [23] (BRISQUE = 37.01). (e) Result of ARCNN [2] (BRISQUE = 33.15). (f) Result of DnCNN [4] (BRISQUE = 33.09). (g) Result of MWCNN [5] (BRISQUE = 37.65). (h) Result of SCENet (BRISQUE = 25.55).
Sensors 19 01939 g008
Figure 9. Comparison of the subjective visual quality for the image from the SIDD dataset, Desk. (a) Captured image using an iPhone 7 with ISO = 200 and exposure = 1/400 (BRISQUE = 46.81). (b) JPEG-coded image with q = 30 (BRISQUE = 28.98). (c) Result of BM3D [22] (BRISQUE = 22.20). (d) Result of SSRQC [23] (BRISQUE = 26.07). (e) Result of ARCNN [2] (BRISQUE = 24.35). (f) Result of DnCNN [4] (BRISQUE = 21.78). (g) Result of MWCNN [5] (BRISQUE = 24.11). (h) Result of SCENet (BRISQUE = 19.70).
Figure 9. Comparison of the subjective visual quality for the image from the SIDD dataset, Desk. (a) Captured image using an iPhone 7 with ISO = 200 and exposure = 1/400 (BRISQUE = 46.81). (b) JPEG-coded image with q = 30 (BRISQUE = 28.98). (c) Result of BM3D [22] (BRISQUE = 22.20). (d) Result of SSRQC [23] (BRISQUE = 26.07). (e) Result of ARCNN [2] (BRISQUE = 24.35). (f) Result of DnCNN [4] (BRISQUE = 21.78). (g) Result of MWCNN [5] (BRISQUE = 24.11). (h) Result of SCENet (BRISQUE = 19.70).
Sensors 19 01939 g009
Table 1. Comparison of the objective quality on classical images in terms of PSNR (dB) and SSIM. Bold values indicate the best scores.
Table 1. Comparison of the objective quality on classical images in terms of PSNR (dB) and SSIM. Bold values indicate the best scores.
MethodNoise LevelMetricCamera ManBarbaraBirdButter FlyHall MonitorLenaMobilePeppers
Degradedq = 10,
peak = 200
PSNR25.8525.1330.8627.9027.9129.8921.8529.39
SSIM0.7370.7530.8050.8360.7710.7770.7500.754
q = 20,
peak = 400
PSNR27.6727.4632.8430.1730.2831.8224.1631.15
SSIM0.7760.8270.8270.8730.8150.8190.8220.797
q = 30,
peak = 600
PSNR28.9729.1734.0231.4131.5632.8425.7332.04
SSIM0.8140.8640.8500.8880.8390.8410.8590.819
BM3D
[22]
q = 10,
peak = 200
PSNR27.2227.1833.1329.9729.8931.2922.8331.53
SSIM0.8090.8050.9010.9130.8580.8020.7820.825
q = 20,
peak = 400
PSNR29.1629.6935.7032.3832.2033.0825.4332.93
SSIM0.8460.8800.9200.9320.8990.8400.8760.842
q = 30,
peak = 600
PSNR30.4731.2137.1433.5933.7534.2926.9533.27
SSIM0.8710.9020.9210.9310.9020.8580.8970.844
SSRQC
[23]
q = 10,
peak = 200
PSNR27.3227.3133.1430.5329.7331.2723.0531.49
SSIM0.8140.8210.8910.9130.8530.8050.7940.818
q = 20,
peak = 400
PSNR28.9729.3235.2832.6532.2032.9425.2932.91
SSIM0.8440.8720.9050.9300.8800.8350.8640.836
q = 30,
peak = 600
PSNR30.1830.9136.3833.6333.3634.2726.7433.61
SSIM0.8680.8980.9150.9360.8910.8500.8920.847
ARCNN
[2]
q = 10,
peak = 200
PSNR27.3426.6333.3331.0430.0131.8823.2231.27
SSIM0.8190.8110.9010.9180.8800.8470.8070.827
q = 20,
peak = 400
PSNR28.8629.1835.1133.2232.4933.5625.7032.57
SSIM0.8380.8730.9060.9310.8980.8720.8680.848
q = 30,
peak = 600
PSNR30.1530.8536.2634.3733.7634.4327.2233.38
SSIM0.8600.9020.9120.9370.9070.8850.8970.860
DnCNN
[4]
q = 10,
peak = 200
PSNR27.8127.1533.5031.1730.5231.9223.6031.55
SSIM0.8330.8170.9030.9190.8860.8440.8090.823
q = 20,
peak = 400
PSNR29.5829.6335.3333.5232.9433.6326.2232.95
SSIM0.8520.8730.9090.9340.9010.8670.8750.839
q = 30,
peak = 600
PSNR30.7531.1836.3534.6233.9534.4427.8633.61
SSIM0.8700.8970.9130.9380.9050.8770.9060.847
MWCNN
[5]
q = 10,
peak = 200
PSNR28.0927.7133.2331.1030.6231.9123.6931.53
SSIM0.8300.8200.8940.9180.8800.8450.8090.821
q = 20,
peak = 400
PSNR29.6829.9234.8533.2632.7433.3526.1532.72
SSIM0.8390.8710.8890.9270.8860.8560.8710.828
q = 30,
peak = 600
PSNR30.9831.5536.0734.5434.0034.3227.7433.49
SSIM0.8640.8980.9000.9310.8970.8700.9000.841
SCENetq = 10,
peak = 200
PSNR28.2427.6734.0831.7231.1332.3723.9032.12
SSIM0.8460.8250.9160.9320.9050.8510.8220.835
q = 20,
peak = 400
PSNR30.2030.3236.4134.2133.8534.2526.5433.68
SSIM0.8800.8870.9350.9510.9300.8790.8930.856
q = 30,
peak = 600
PSNR31.3631.8437.5835.3735.1035.1828.1734.43
SSIM0.8940.9100.9390.9550.9360.8910.9200.866
Table 2. Comparison of the objective quality on the LIVE1 database in terms of average PSNR (dB) and SSIM. Bold values indicate the best scores.
Table 2. Comparison of the objective quality on the LIVE1 database in terms of average PSNR (dB) and SSIM. Bold values indicate the best scores.
Noise LevelMetricDegradedBM3D [22]SSRQC [23]ARCNN [2]DnCNN [4]MWCNN [5]SCENet
q = 10,
peak = 200
PSNR27.0127.8528.0328.1628.6028.6428.97
SSIM0.7300.7610.7620.7740.7970.7950.805
q = 20,
peak = 400
PSNR28.9930.1529.8830.2230.6030.4731.08
SSIM0.7940.8330.8350.8430.8500.8410.863
q = 30,
peak = 600
PSNR30.1831.3831.3031.3531.5831.6332.34
SSIM0.8270.8610.8590.8680.8730.8710.890

Share and Cite

MDPI and ACS Style

Yoo, S.B.; Han, M. SCENet: Secondary Domain Intercorrelation Enhanced Network for Alleviating Compressed Poisson Noises. Sensors 2019, 19, 1939. https://doi.org/10.3390/s19081939

AMA Style

Yoo SB, Han M. SCENet: Secondary Domain Intercorrelation Enhanced Network for Alleviating Compressed Poisson Noises. Sensors. 2019; 19(8):1939. https://doi.org/10.3390/s19081939

Chicago/Turabian Style

Yoo, Seok Bong, and Mikyong Han. 2019. "SCENet: Secondary Domain Intercorrelation Enhanced Network for Alleviating Compressed Poisson Noises" Sensors 19, no. 8: 1939. https://doi.org/10.3390/s19081939

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop