Next Article in Journal
Modeling and Analysis of Social Phenomena: Challenges and Possible Research Directions
Previous Article in Journal
Gravity, Quantum Fields and Quantum Information: Problems with Classical Channel and Stochastic Theories
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Search Dense Connection Module for Super-Resolution

Key Laboratory of Knowledge Engineering with Big Data, Ministry of Education, School of Computer Science and Information Engineering, Hefei University of Technology, Hefei 230601, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(4), 489; https://doi.org/10.3390/e24040489
Submission received: 7 March 2022 / Revised: 28 March 2022 / Accepted: 29 March 2022 / Published: 31 March 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
The development of display technology has continuously increased the requirements for image resolution. However, the imaging systems of many cameras are limited by their physical conditions, and the image resolution is often restrictive. Recently, several models based on deep convolutional neural network (CNN) have gained significant performance for image super-resolution (SR), while extensive memory consumption and computation overhead hinder practical applications. For this purpose, we present a lightweight network that automatically searches dense connection (ASDCN) for image super-resolution (SR), which effectively reduces redundancy in dense connection and focuses on more valuable features. We employ neural architecture search (NAS) to model the searching of dense connections. Qualitative and quantitative experiments on five public datasets show that our derived model achieves superior performance over the state-of-the-art models.

1. Introduction

Since it is difficult for the current visual effects of images to meet people’s needs, single image super-resolution (SISR) [1] and its related technologies have attracted widespread attention [2]. SISR is a low-level computer vision task for reconstructing high-resolution images from low-resolution images. Due to the rapid development of deep convolutional neural networks (CNNs), deep CNN-based approaches have gained better reconstruction results against traditional methods in the field of SISR [3,4].
Super-resolution convolutional neural network (SRCNN) [5] was the pioneer of deep CNN in super-resolution (SR) problems, and an end-to-end nonlinear mapping through only a three-layer convolutional network was established. Since then, numerous CNN-based algorithms have emerged and made remarkable progress. Very deep convolutional networks super-resolution (VDSR) [6] deepen the network to 20 layers by exploiting residual learning [7], which alleviated the training difficulty. Deep recursive residual network (DRRN) [8] equipped recursive blocks to obtain promising results with deeper network structures. However, these methods used the interpolated low-resolution (LR) images as input in the network, which undoubtedly led to computational and time burdens. Shi et al. [9] devised an efficient sub-pixel convolution to tackle this problem, which directly extracted feature maps from the LR images. Subsequently, Lim et al. [10] extended the depth and width of the network, and achieved significant performance gains by eliminating batch normalization modules in residual networks.
As mentioned above, deep CNN-based SISR architectures have yielded great success but have not fully leveraged the multiscale representation and the intermediate features [11]. Then, Lan et al. [12] explored a model that combines multi-scale residuals with attention mechanism, which can not only extract multi-level features, but also exploits the discriminative information of different channels. A one-shot aggregation network (OAN) [13] employed diverse features with multiple receptive fields by aggregating all previous features into subsequent layers. Inspired by DenseNet [14], Zhang et al. [15] further integrated the dense structure and the residual structure to form a residual dense network (RDN) to exploit the hierarchical features. DenseNet proposed dense connectivity to improve computational efficiency via encouraging feature reuse. However, Huang et al. [16] pointed out that abusing dense connectivity led to redundancies. Each layer did not need to receive information from all the previous layers. This could take up large amounts of memory, which largely restricted the applications on resource-constrained mobile platforms. Hence, Huang et al. [16] introduced CondenseNet, which adopted learned group convolution (LGC) to prune these redundant connections. This produced an efficient, lightweight dense connection network.
Motivated by this, we remove the less important connections from a different viewpoint in this paper. We present a novel method automatic search dense connection network (ASDCN), which utilizes gradient-based neural architecture search (NAS) models [17] to cut superfluous connections automatically. Our entire network has two training procedures. In the first stage, our network selects the right connections through efficient dense connection search. In the second stage, the appropriate structure is trained according to the architecture parameters learned in the first stage. We observe that connection values greater than 0.1 contribute significantly to network performance, and these connections are considered essential to derive the final architecture. Our search space is only for the densely connected joint patterns, without search operations, which can effectively get rid of redundancy. Meanwhile, our proposed method achieves promising results with few parameters. The main contributions of this paper are summarized as follows:
  • We introduce a novel lightweight ASDCN model for single image super-resolution, selecting key connection paths effectively, and suppressing redundant information.
  • We equip a softmax function to relax the dense connection paths into a continuous space and integrate the architecture search into the model for training. According to the weights of the paths, the appropriate connections are screened out. Selecting the essential features from intermediate layers enables the network to be more compact and efficient.
  • Comprehensive experiments on five public benchmark datasets have demonstrated that our derived model achieves comparable performance to the most advanced methods. Our proposed method strikes a trade-off between reconstruction results and model sizes.

2. Related Work

2.1. Deep CNN-Based Super-Resolution

Due to the emergence of large-scale labeled data and the rapid improvements of GPU [18], CNN-based methods have developed rapidly and obtained state-of-the-art results in a variety of studies [19,20].
Dong et al. [5] first explored a three-layer CNN for SR reconstruction and made a huge improvement compared to traditional methods. Thanks to residual learning, the vanishing gradient problem in deep networks was alleviated. Then, a deeply-recursive convolutional network (DRCN) [21] deployed a deep network by combining residual and recursive ideas, increasing the receptive fields and improving performance. Multi-scale deep super-resolution (MDSR) [10] was devised for simultaneous multi-scale image learning at different magnifications, giving the network sufficient mapping capability and winning the NTIRE2017 challenge. A densely residual Laplacian network (DRLN) [22] utilized dense connections between residual blocks to promote Laplacian attention to assign weights at different scales, resulting in considerable performance gains. Although these networks produce state-of-the-art results, they require large amounts of memory, powerful computing ability and long inference times, and are not suitable for deployment on mobile devices with constrained resources.
Some researchers focused on developing lightweight, but efficient models for SISR [23] without reducing accuracy. A cascading residual network (CARN) [24] constructed a lightweight cascaded residual network through a cascaded scheme with group convolution. Not only did it maintain the most advanced performance but it was faster. An adaptive weighted super-Resolution network (AWSRN) [25] provided an adaptive weighted residual unit to automatically calculate the total residual and initial mapping parameters, achieving better reconstruction quality with lower complexity. In addition, Tian et al. [26] developed a coarse-to-fine CNN for SISR (CFSRCNN), which cascaded multiple hierarchical features to prevent possible training instability and performance degradation, and remarkably improved the computational efficiency. These efficient CNN based models are hand-crafted super-resolution networks. The following section will introduce the NAS-based approaches for SR to achieve optimal performance in an automated manner.

2.2. Neural Architecture Search

Neural architecture search (NAS) is an algorithm that automatically learns the appropriate deep neural structure for a specific task with minimal human involvement. The pioneering work of NAS was conducted by [27], who employed the reinforcement learning (RL) method to produce higher accuracy in image classification tasks. Subsequently, evolutionary algorithms [28,29] were introduced to solve NAS problems and achieved considerable classification accuracy on a large scale. Nevertheless, these methods bear hundreds of GPU days. Hence, researchers began to wonder how to reduce the amount of computation and speed up the search for neural structures. An example of this is efficient-NAS (ENAS) [30], which proposed a weight-sharing strategy to improve search efficiency. Compared with NAS, ENAS [30] could shorten GPU computing time by more than 1000 times. Differentiable architecture search (DARTS) [31], another variant of NAS, relaxed a given discrete search space into a continuous space by conducting architecture searches in a differentiable way, and is orders of magnitude faster than the most advanced non-differentiable algorithms. In this paper, we adopt the DARTS algorithm to search the connection pattern of dense connections.
Most of the existing super-resolution models are designed manually and are difficult to compress or fine-tune. At the same time, the neural architecture search algorithm has been highly influential in classification tasks. According to this trend, Chu et al. [32,33] presented fast, accurate, and lightweight SR (FALSR) and multi-objective reinforced evolution in mobile NAS (MoreMNAS), which dealt with super-resolution utilizing a multi-objective method. FALSR-C [32] (a more lightweight version of FALSR) indicated that unwanted features from lower layers could cause problems for high layers to reconstruct SR results. Song et al. [34] built three efficient residual dense blocks to search lightweight SR networks with the evolutionary approach. In addition, these based NAS methods for SR produced brilliant results.

3. Proposed Method

3.1. Network Architecture

In this section, we introduce our proposed approach. With RDN [15] architecture as the backbone, our proposed network named “automatic search dense connection” (ASDCN) mainly consists of three parts: shallow feature extraction, a nonlinear mapping module with several automatic search dense connection blocks (ASDCBs), and a reconstruction part, as shown in Figure 1. I L R and I S R represent the input raw image and the corresponding high-resolution output of the network, respectively.
Firstly, we leverage one convolution layer to extract low-level features from the original input image, which can be denoted as
X 0 = f e x t ( I L R )
where f e x t is a convolution layer with a kernel size of 3 × 3 to extract the primitive features from the LR image I L R .
Afterwards, the output feature maps x 0 are fed into the following nonlinear mapping module consisting of a series of stacked multiple ASDCBs to gradually obtain the hierarchical features, which can be represented as
X t = f B t ( X t 1 ) = f B t ( f B t 1 ( f B 0 ( X 0 ) ) )
where X t 1 and X t are the input feature maps and output feature maps of the t-th ASDCB, respectively. After obtaining the multi-level powerful feature representations, the generated features are concatenated through global feature fusion, which can be denoted as
X o u t p u t = F o u t p u t ( C a t ( X 1 , , X n ) )
where F o u t p u t denotes a convolution layer with a kernel size of 1 × 1 . Then, the features X o u t p u t are up-sampled to the HR image size via an upscaling module. Moreover, two up-sampling modules are required when the scaling factor is × 4 . The upscaling module is made up of the nearest neighbor (NN) layer and a pixel attention [35] with two convolution layers interleaved. Each convolution layer is followed by an LReLU [36] activation function while the pixel attention layer contains a 1 × 1 convolution layer and a sigmoid function. Finally, the interpolated I L R is added to the upsampled X o u t p u t by global residual connection to obtain the final predicted SR image of the network as follows:
I S R = f r e c ( X o u t p u t ) + f u p ( I L R )
where f r e c represents the reconstruction module, and f u p stands for the bilinear interpolation.
Following previous works [15,24], our network is optimized by L 1 loss function to measure the difference between the predicted SR image and the ground truth HR image. Given a training set { I L R i , I H R i } i = 1 N , where N is the number of LR-HR training patches, the loss function of our SR network can be expressed as
L ( Θ ) = 1 N k = 1 N | | H A S D C N ( I L R i ) I H R i | | 1
where H A S D C N denotes our proposed model while Θ indicates the parameters set within it.

3.2. Automatic Search Dense Connection Module

This section describes how to search for dense connections using the gradient-based NAS method. As is common knowldge, DenseNet allows feature maps from all previous layers entering into the subsequent layers to make the most use of the features. However, there remain some redundant connections which will affect the efficiency of the network. It is difficult to determine which remaining features are unnecessary. To this end, we devise an adaptive structure to prune unimportant connections while retaining useful ones during training.
Figure 2 shows our proposed automatic search dense connection block (ASDCB). Our search space is composed of dense connections between distinct layers of each block. The key idea of the proposed method is to relax the discrete densely connected space into a continuous representation, which allows us to choose the candidate paths with significant contributions in a differentiable manner. To this end, we exploit a softmax function for the continuous relaxation of the search space. We assign an initial probability parameter to each path of dense connection between various layers in one block. During the search process, the probability parameters are optimized. We further sort all the candidate paths, which can help us to screen out the path with a more outstanding contribution and eliminate the path with less of a contribution. By doing so, superior candidate architectures can be searched for further experiments to obtain better results.
We relax dense connections into continuous representations and assign an architecture parameter α to each output path of the layer. Let o be the set of candidate connection paths, and α o ( i , j ) be the weight of each output path of the layer. We employ a softmax function to compute the probability of each input path over all paths in one layer as follows:
P ( α o ( i , j ) ) = exp ( α o ( i , j ) ) o O exp ( α o ( i , j ) )
The output of each layer is computed based on all of its previous layer in one block, and can be expressed as
x n + 1 = o O P ( α o ( i , j ) ) · o ( x n )
where x i represents the input tensor of layer n, and o stands for the convolution operation. Hence, the architecture search can be treated as an optimization problem for a set of continuous variables α = α o ( i , j ) .
Through an automatic searching strategy, the connection paths that contribute most are selected, while other paths are discarded. Then, the final architecture is derived from the learned parameters. The search process is described in the next section.

3.3. Search Procedure

Based on the continuous relaxation of the search space, we can leverage the gradient descent strategy to optimize the architecture parameters and network weights jointly. Let α be the parameters of the proposed module and ω be the parameters of the whole network, and the training process can be described as:
min α L v a l ( ω * ( α ) , α )
s . t . ω * ( α ) = arg min ω L t r a i n ( ω , α )
where L t r a i n and L v a l indicate the training and validation loss, respectively.
We aim to jointly optimize the architecture parameters α and weights ω of the network, so that the architecture finds the minimum training and validation loss. First, we optimize the network weights ω by descending ω L t r a i n ( ω , α ) for enough epochs to warm up on the training dataset. After warming up ω , we update the architecture parameters by descending ω L t r a i n ( ω * ( α ) , α ) into validation datasets. The architecture parameters α are randomly initialized. Then, P ( α o ( i , j ) ) is defined as the importance of all input paths. A particular layer can be calculated by Formula (6) to determine which paths are retained. We set a threshold of 0.1 (the paths will be discarded when P ( α o ( i , j ) ) < 0.1 ). According to these learned parameters, we choose the most appropriate candidate paths to derive the final architecture for experiments. The whole search procedure is shown in Algorithm 1.
Our proposed algorithm shows that dense connections are not always the best way to transmit information. Features with small contributions from lower layers can cause problems in reconstructing super-resolution results at the high-level layer. We discard input paths that contribute little to each layer in the block during the search process, while reserving input paths with high weights. Compared with the pruning weights in a pre-trained network, our method is lighter and more efficient. It not only restricts the front-end redundancy of DenseNet and reduces the number of parameters, but also achieves competitive performance.
Algorithm 1 Training process.
Require: 
epoch, architecture parameter α ( i , j )
  1:
         network weight ω
  2:
initialize α and ω
  3:
for epoch ≤ 20 do
  4:
   update ω
  5:
end for
  6:
for not converged do
  7:
   update 1 α α L v a l ( ω ε ω L t r a i n ( ω , α ) , α )
  8:
   update 2 ω ω L t r a i n ( ω , α )
  9:
end for
10:
Derive the final architecture and retrain.

4. Experiments

4.1. Datasets and Metrics

In order to make fair comparisons with the state-of-the-art SR algorithms, we follow previous works. DIV2K [37] is a recent high-resolution dataset, which includes 800 training images, 100 validation images, and 100 test images. We adopt 800 pairs of LR and HR training images from DIV2K to train our model, and the LR images are obtained via the bicubic downsampling of the corresponding HR images. In addition, Set5 [38] is adopted for validation after each epoch. In the testing phase, we employ several public benchmark datasets (Set14 [39], BSD100 [40], Urban100 [41], and Manga109 [42]) to evaluate the performance of our proposed algorithm under three upscaling factors ( × 2 , × 3 , and × 4 ). The peak signal-to-noise ratio (PSNR) and the structural similarity index (SSIM) [43] on the Y channel of transformed YCbCr space are treated as quantitative evaluation metrics.
Given a ground-truth image I H R and a predicted image I S R , the P S N R is formulated as:
P S N R ( I H R , I S R ) = 10 l o g 10 ( M a x I 2 M S E )
where
M S E = 1 H × W i = 1 H j = 1 W ( I H R ( i , j ) I S R ( i , j ) ) 2
M a x I is the maximum pixel value of an image. H and W are the height and width of a image, respectively. S S I M is defined as:
S S I M ( I H R , I S R ) = l ( I H R , I S R ) c ( I H R , I S R ) s ( I H R , I S R ) ,
where
l ( I H R , I S R ) = 2 μ I H R μ I S R + C 1 μ I H R 2 + μ I S R 2 + C 1 c ( I H R , I S R ) = 2 σ I H R σ I S R + C 2 σ I H R 2 + σ I S R 2 + C 2 s ( I H R , I S R ) = 2 σ I H R I S R + C 3 σ I H R σ I S R + C 3
μ I H R , σ I H R , and σ I H R I S R are the mean values, variance, and covariance of an image, respectively. C 1 , C 2 and C 3 are set to positive constants to avoid instability when the denominator is close to zero.
Moreover, we calculate Multi-Adds and the number of parameters to assess the complexity of our model. Multi-Adds are computed on HR images with a spatial resolution of 720p at all scales.

4.2. Implementation Details

The whole training process is split into the searching and retraining phases. The MATLAB function is applied to the bicubic downsampling of the counterpart HR images to obtain the corresponding LR images. We randomly crop image patches with the size of 96 × 96 for the searching phase ( 144 × 144 for the retraining phase) from the LR images, and 16 patches are utilized as input for each training minibatch. Data augmentation is conducted by random rotations of 90 , 180 , 270 , and horizontal flips for each training iteration. The searching stage and retraining stage contain 200 and 1000 epochs, respectively. We update only network weights for the first 20 epochs in the searching stage. Then, the architecture parameters are updated using the early stop strategy. We set 1000 iterations as an epoch.
In our model, except for feature fusion parts equipped with 1 × 1 convolution, the other parts all use a 3 × 3 convolution layer. Furthermore, we employ a padding strategy to keep the size of the output feature maps the same for each layer. Our ASDCN model consists of ten automatic search dense connection blocks (ASDCBs). Each block has six 3 × 3 convolution layers and a 1 × 1 convolution layer. A convolution layer with a kernel size of 1 × 1 is employed to match the channels and preserve more useful information. The channel of the intermediate layer within each block is fixed to 16.
Our network is optimized using the Adam [44] optimizer by setting β 1 = 0.9 and β 1 = 0.999 . We leverage cosine annealing to reduce the learning rate. The maximum learning rate is initialized as 10 4 , and the target minimum learning rate is fixed as 10 6 . The learning rate of the architecture parameter is set to 0.002 . Additionally, different from reference [45] using dynamic differential evolution, for the hyperparameters of network search, the best ones are selected through repeated experiments on the search process. For the hyperparameters of the network structure, several combinations of the number of blocks and convolutional layers are repeatedly tested to choose the optimal ones under a specific parameter amount. The other hyperparameters are provided by the reference image super-resolution methods without trial and error. Our model is trained using the PyTorch framework with an NVIDIA RTX 2080Ti GPU.

4.3. Ablation Study

4.3.1. Comparison with RDN with the Same Setup

In this section, we make a comparison with RDN [15] under the same setup. We employ the same training dataset to train these two models. The two models have six blocks, and the channel number of intermediate layers is set to 16. The experimental results with a scaling factor of × 2 in five available datasets are shown in Table 1. It can be seen that our proposed algorithm is slightly better than RDN with PSNR and SSIM, but Multi-Adds and the number of parameters are significantly fewer than RDN. This indicates that dense connections are not always the best way to transmit information. There are still redundant connections, and it is not necessary to feed each previous layer into the next layer. Our designed strategy for automatically searching dense connection patterns can selectively use the essential features from previous layers, which can reduce redundancy and improve network efficiency. It is further shown that the adaptive dense network improvement method can obtain a lightweight model with comparable performance.

4.3.2. Searched Architectures

This section shows the internal connection patterns for the first, fifth, and tenth ASDCB architecture. It is clear from Figure 3 that not every intermediate layer accepts the output of each previous layer. Thus, the adaptive selection of connection paths can effectively reduce redundancy and boost the efficiency of the network without degrading performance.

4.4. Comparison with State-of-the-Art Methods

To demonstrate the effectiveness of our proposed architecture, we compared ASDCN with other state-of-the-art models, including SRCNN [5], VDSR [6], DRCN [21], Laplacian pyramid super-resolution network (LapSRN) [3], CARN-M [24], MoreMNAS-A [33], FALSR-C [32], AWSRN-S [25], efficient super-resolution network (ESRN-V) [34], multiscale attention dual network (MADNet-L1) [12], OAN-S [13], and weighted multi-scale residual network (WMRN) [4]. These are lightweight models within the number of parameters 2.0   M or Multi-Adds 100   G . Parameters (space complexity) and Multi-Adds (time complexity) are used to reflect the complexity of our model. The quantitative comparisons for × 2 , × 3 , and × 4 are depicted in Table 2 on the five datasets.
As can be seen from Table 2, the proposed model is superior to the most advanced models at different scaling factors with less than 2 M parameters. Under comparable computational complexity, our ASDCN achieves higher PSNR values than hand-designed CARN-M. Compared with the manually constructed WMRN, our derived architecture obtains a better reconstruction result, but the number of parameters and Multi-Adds are reduced by about 53 % . Compared with MADNet-L1, ASDCN achieves higher reconstruction accuracy with fewer parameters and Multi-Add. Moreover, our searched model also outperforms the three most advanced NAS-based approaches (FALSR-C, ESRN-V, and MoreMNAS-A) for × 2 SR on all the benchmark datasets. Specifically, MoreMNAS-A has three times as large as ours in terms of parameters and Multi-Adds, but our PSNR value has a considerable margin of 0.63 dB PSNR that of Urban100.
In addition, Table 2 also provides the complexity (the amount of parameters and Muti-Adds) of the different models for a more intuitive comparison. The parameters and Muti-Adds of SRCNN and VDSR do not change at the all scales because the bicubic interpolation images are required as input, and other methods have inconsistent changes. Since our model has relatively few parameters and Multi-Adds is also not high, it is a lightweight model.
Figure 4 further compares the number of parameters and performance of different approaches. The results show that our method exceeds other methods in both parameters and performance, which fully proves that we have achieved a better balance between model size and performance.
In addition to quantitative evaluation, we also visually compare our model with other models. Figure 5 shows the subjective visual quality on three datasets with × 4 the upscaling factor. For the “img_024” and “img_076” from Urban100, only our method can restore the correct lines and suppress the distortions, whereas other methods cannot reconstruct the proper structure. For “HighschoolKimengumi_vol20” from Manga109, our network has more precise texture information and edges with less blurring and artifacts. Furthermore, for “0823” from BSD100, we can also notice that our derived architecture produces the best reconstruction effect, whereas VDSR has checkerboard artifacts, and CARN-M has more blurring and noise.

4.5. Visualization on Real-World Images

To verify the effectiveness and robustness of the proposed algorithm, we further compare it with other methods on real-world images. There is no high-quality ground truth in these cases, and the degradation model is unknown.
Figure 6 shows that our method can precisely recover more image details and more apparent contours. The better perceptual quality further indicates that our derived architecture can search for more convincing SR models.

5. Conclusions

In this paper, we propose a framework for automatically searching dense connection modules for single image super-resolution. The NAS-based method used to search dense connection paths can adaptively select the key connection paths and effectively reduce the redundant information of the network. Moreover, it is more efficient than manual pruning. The lightweight image super-resolution is realized by efficient residual dense connection blocks and multi-layer information fusion. Extensive quantitative and qualitative experiments demonstrate that our derived model is superior to most state-of-the-art approaches with comparable parameters and Multi-Adds.
Our model only searches the intermediate nodes of the dense block, and only searches which pre-nodes need to be used and which can be discarded. Our future work will extend to search the operation of the model, and even search the block level of the entire model synchronously. Of course, we can also study non-NAS methods later, such as permuting and combining more basic operations (multiplication, addition, etc.).

Author Contributions

Conceptualization, H.Z. and S.Z.; methodology, H.Z. and S.Z.; formal analysis, H.Z., G.C. and Y.Z.; writing—original draft preparation, H.Z.; writing—review and editing, S.Z., G.C. and Z.D.; visualization, H.Z.; supervision, S.Z.; project administration, S.Z. and G.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Anhui Province R & D Key Project under Grant No. JD2019XKJH0029 and Hefei Municipal Natural Science Foundation under Grant No. 2021008.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Guo, T.; Dai, T.; Liu, L.; Zhu, Z.; Xia, S.-T. S2A:Scale Attention-Aware Networks for Video Super-Resolution. Entropy 2021, 23, 1398. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, K.; Liang, J.; Gool, L.V.; Timofte, R. Designing a practical degradation model for deep blind image super-resolution. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Online, 11–17 October 2021; pp. 4771–4780. [Google Scholar]
  3. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep Laplacian pyramid networks for fast and accurate super-resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 624–632. [Google Scholar]
  4. Sun, L.; Liu, Z.; Sun, X.; Liu, L.; Lan, R.; Luo, X. Lightweight Image Super-Resolution via Weighted Multi-Scale Residual Network. IEEE/CAA J. Autom. Sin. 2021, 8, 1271–1280. [Google Scholar] [CrossRef]
  5. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the European Conference on Computer Vision–ECCV 2014, Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar]
  6. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1646–1654. [Google Scholar]
  7. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  8. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2790–2798. [Google Scholar]
  9. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1874–1883. [Google Scholar]
  10. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced deep residual networks for single image super-resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  11. Li, Z.; Wang, C.; Wang, J.; Ying, S.; Shi, J. Lightweight adaptive weighted network for single image super-resolution. Comput. Vis. Image Underst. 2021, 211, 103254. [Google Scholar] [CrossRef]
  12. Lan, R.; Sun, L.; Liu, Z.; Lu, H.; Pang, C.; Luo, X. MADNet: A Fast and Lightweight Network for Single-Image Super Resolution. IEEE Trans. Cybern. 2021, 51, 1443–1453. [Google Scholar] [CrossRef] [PubMed]
  13. Tang, R.; Chen, L.; Zou, Y.; Lai, Z.; Albertini, M.K.; Yang, X. Lightweight network with one-shot aggregation for image super-resolution. J. Real-Time Image Process. 2021, 18, 1275–1284. [Google Scholar] [CrossRef]
  14. Huang, G.; Liu, Z.; Maaten, L.V.D.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
  15. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  16. Huang, G.; Liu, S.; der Maaten, L.V.; Weinberger, K.Q. Condensenet: An efficient densenet using learned group convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  17. Baker, B.; Gupta, O.; Naik, N.; Raskar, R. Designing neural network architectures using reinforcement learning. arXiv 2016, arXiv:1611.02167. [Google Scholar]
  18. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2014, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  19. Quan, R.; Yu, X.; Liang, Y.; Weinberger, Y. Removing raindrops and rain streaks in one go. In Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Online, 19–25 June 2021; pp. 9147–9156. [Google Scholar]
  20. Peng, H.; Chen, X.; Zhao, J. Densely connected convolutional networks. In Proceedings of the 2020 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 2012–2020. [Google Scholar]
  21. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-recursive convolutional network for image super-resolution. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1637–1645. [Google Scholar]
  22. Anwar, S.; Barnes, N. Densely Residual Laplacian Super-Resolution. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1192–1204. [Google Scholar] [CrossRef] [PubMed]
  23. Hui, Z.; Wang, X.; Gao, X. Fast and accurate single image superresolution via information distillation network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 723–731. [Google Scholar]
  24. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  25. Wang, C.; Li, Z.; Shi, J. Lightweight Image Super-Resolution with Adaptive Weighted Learning Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  26. Tian, C.; Xu, Y.; Zuo, W.; Zhang, B.; Fei, L.; Lin, C.W. Coarse-to-fine cnn for image super-resolution. IEEE Trans. Multimed. 2020, 23, 1489–1502. [Google Scholar] [CrossRef]
  27. Zoph, B.; Le, Q.V. Neural architecture search with reinforcement learning. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  28. Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized evolution for image classifier architecture search. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 4780–4789. [Google Scholar]
  29. Zhang, T.; Lei, C.; Zhang, Z.; Meng, X.-B.; Chen, C.L.P. AS-NAS: Adaptive Scalable Neural Architecture Search With Reinforced Evolutionary Algorithm for Deep Learning. IEEE Trans. Evol. Comput. 2021, 25, 830–841. [Google Scholar] [CrossRef]
  30. Pham, H.; Guan, M.Y.; Zoph, B.; Le, Q.V.; Dean, J. Efficient neural architecture search via parameter sharing. In Proceedings of the 32nd International Conference on Machine Learning (ICML), Vienna, Austria, 25–31 July 2018. [Google Scholar]
  31. Liu, H.; Simonyan, K.; Yang, Y. DARTS: Differentiable architecture search. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  32. Chu, X.; Zhang, B.; Ma, H.; Xu, R.; Li, J.; Li, Q. Fast, accurate and lightweight super-resolution with neural architecture search. In Proceedings of the Conference: 2020 25th International Conference on Pattern Recognition (ICPR), Online, 10–15 January 2020. [Google Scholar]
  33. Chu, X.; Zhang, B.; Xu, R.; Ma, H. Multi-objective reinforced evolution in mobile neural architecture search. In Proceedings of the European Conference on Computer Vision (ECCV), Online, 23–28 August 2020. [Google Scholar]
  34. Song, D.; Xu, C.; Jia, X.; Chen, Y.; Xu, C.; Wang, Y. Efficient residual dense block search for image super-resolution. In Proceedings of the The Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI), New York, NY, USA, 7–12 February 2020. [Google Scholar]
  35. Peng, H.; Chen, X.; Zhao, J. Residual pixel attention network for spectral reconstruction from RGB images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  36. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the ICML Workshop on Deep Learning for Audio, Speech and Language Processing, Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  37. Agustsson, E.; Timofte, R. Ntire 2017 challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA, 21–26 June 2017. [Google Scholar]
  38. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the British Machine Vision Conference, Surrey, UK, 3–7 September 2012; pp. 135.1–135.10. [Google Scholar]
  39. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Curves and Surfaces; Boissonnat, J.-D., Chenin, P., Cohen, A., Gout, C., Lyche, T., Mazure, M.-L., Schumaker, L., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 711–730. [Google Scholar]
  40. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the 8th International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  41. Huang, J.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  42. Matsui, Y.; Ito, K.; Aramaki, Y.; Fujimoto, A.; Ogawa, T.; Yamasaki, T.; Aizawa, K. Sketch-based manga retrieval using manga109 dataset. Multimed. Tools Appl. 2017, 76, 21811–21838. [Google Scholar] [CrossRef] [Green Version]
  43. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  45. Singh, D.; Kaur, M.; Jabarulla, M.Y.; Kumar, V.; Lee, H.N. Evolving fusion-based visibility restoration model for hazy remote sensing images using dynamic differential evolution. IEEE Trans. Geosci. Remote Sens. 2022, 15, 5765. [Google Scholar] [CrossRef]
Figure 1. An overview of our ASDCN architecture.
Figure 1. An overview of our ASDCN architecture.
Entropy 24 00489 g001
Figure 2. The automatic dense connection module.
Figure 2. The automatic dense connection module.
Entropy 24 00489 g002
Figure 3. The searched architectures of different blocks.
Figure 3. The searched architectures of different blocks.
Entropy 24 00489 g003aEntropy 24 00489 g003b
Figure 4. Comparison of the performance and parameters between our ASDCN model and other models on Urban100 with a scale factor of 2.
Figure 4. Comparison of the performance and parameters between our ASDCN model and other models on Urban100 with a scale factor of 2.
Entropy 24 00489 g004
Figure 5. Visual comparison of × 4 super-resolution images on the Urban100, BSD100, and Manga109 datasets. The best results are highlighted by red.
Figure 5. Visual comparison of × 4 super-resolution images on the Urban100, BSD100, and Manga109 datasets. The best results are highlighted by red.
Entropy 24 00489 g005
Figure 6. Visual comparison with a scale factor × 4 on real-world images.
Figure 6. Visual comparison with a scale factor × 4 on real-world images.
Entropy 24 00489 g006
Table 1. Comparison with RDN on five benchmark datasets. All the other settings are strictly the same. The best performance is highlighted by red.
Table 1. Comparison with RDN on five benchmark datasets. All the other settings are strictly the same. The best performance is highlighted by red.
MethodParamsMulti-AddsSet5Set10BSD100Urban100Manga109
(K)(G)PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
RDN520119.937.90/0.960133.45/0.916532.12/0.899031.87/0.925938.28/0.9764
ASDCN (ours)36483.837.91/0.960333.48/0.917632.12/0.899031.87/0.926138.30/0.9765
Table 2. Quantitative results of several state-of-the-art SR models at scaling factors of × 2 , × 3 and × 4 (average PSNR/SSIM). The best performance is highlighted by red.
Table 2. Quantitative results of several state-of-the-art SR models at scaling factors of × 2 , × 3 and × 4 (average PSNR/SSIM). The best performance is highlighted by red.
MethodScaleParamsMulti-AddsSet5Set10BSD100Urban100Manga109
(K)(G)PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
SRCNN [5]×25752.736.66/0.954232.42/0.906331.36/0.887929.50/0.894635.60/0.9663
VDSR [6]×2665612.637.53/0.958733.03/0.912431.90/0.896030.76/0.914037.22/0.9729
LapSRN [3]×281329.937.52/0.959033.08/0.913031.80/0.895030.41/0.910037.27/0.9740
IDN [23]×2590174.137.83/0.960033.30/0.914832.08/0.895031.27/0.9196-
CARN-M [24]×241291.237.53/0.958333.26/0.914131.92/0.896031.23/0.9193-
MoreMNAS-A [33]×21039238.637.63/0.958433.23/0.913831.95/0.896131.24/0.9187-
FALSR-C [32]×240893.737.66/0.958633.26/0.914031.96/0.896531.24/0.9187-
AWSRN-S [25]×239791.237.75/0.959633.31/0.915132.00/0.897431.39/0.920737.90/0.9755
ESRN-V [34]×232473.437.85/0.960033.42/0.916132.10/0.898731.79/0.9248-
MADNet-L1 [12]×2878187.137.85/0.960033.38/0.916132.04/0.897931.62/0.9233-
OAN-S [13]×2450104.937.85/0.960033.41/0.916232.06/0.898331.61/0.923038.16/0.9761
WMRN [4]×245210337.83/0.959933.41/0.916232.08/0.898431.68/0.924138.27/0.9763
ASDCN(ours)×236483.837.91/0.960333.48/0.917632.12/0.899031.87/0.926138.30/0.9765
SRCNN [5]×35752.732.75/0.909029.28/0.820928.41/0.786326.24/0.798930.59/0.9107
VDSR [6]×3665612.633.66/0.921329.77/0.831428.82/0.797627.14/0.827932.01/0.9310
CARN-M [24]×341246.133.99/0.923630.08/0.836728.91/0.800026.86/0.8263-
IDN [23]×2590105.634.11/0.925329.99/0.835428.95/0.801327.42/0.8359-
AWSRN-S [25]×344748.634.02/0.924030.09/0.837628.92/0.800927.57/0.839132.82/0.9393
ESRN-V [34]×332436.234.23/0.926230.27/0.840029.03/0.803927.95/0.8481-
MADNet-L1 [12]×393088.434.16/0.925330.21/0.839828.98/0.802327.77/0.8439-
OAN-S [13]×349051.234.17/0.925530.20/0.839528.99/0.802327.80/0.843833.06/0.9144
WMRN [4]×35565734.11/0.925130.17/0.839028.98/0.802127.80/0.844833.07/0.9413
ASDCN(ours)×336437.2834.27/0.926630.27/0.841329.06/0.804128.03/0.849933.28/0.9430
SRCNN [5]×45752.730.48/0.862827.49/0.750326.90/0.710124.52/0.722127.66/0.8505
VDSR [6]×4665612.631.35/0.883828.01/0.767427.29/0.725125.18/0.752428.83/0.8809
LapSRN [3]×4813149.431.54/0.885028.19/0.772027.32/0.728025.21/0.756029.09/0.8845
IDN [23]×459081.831.82/0.890328.25/0.773027.41/0.729725.41/0.7632-
CARN-M [24]×441232.531.92/0.890328.42/0.776227.44/0.730425.63/0.7688-
AWSRN-S [25]×458833.731.77/0.889328.35/0.776127.41/0.730425.56/0.767829.74/0.8982
ESRN-V [34]×432420.731.99/0.891928.49/0.777927.50/0.733125.87/0.7782-
MADNet-L1 [12]×4100254.131.95/0.891728.44/0.778027.47/0.732725.76/0.7746-
OAN-S [13]×452042.531.99/0.892628.49/0.797527.49/0.733225.81/0.776030.10/0.9036
WMRN [4]×453645.732.00/0.895228.47/0.778627.49/0.732825.89/0.778930.11/0.9040
ASDCN(ours)×437521.5932.06/0.893728.53/0.780627.54/0.735125.98/0.783130.23/0.9063
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zang, H.; Cheng, G.; Duan, Z.; Zhao, Y.; Zhan, S. Automatic Search Dense Connection Module for Super-Resolution. Entropy 2022, 24, 489. https://doi.org/10.3390/e24040489

AMA Style

Zang H, Cheng G, Duan Z, Zhao Y, Zhan S. Automatic Search Dense Connection Module for Super-Resolution. Entropy. 2022; 24(4):489. https://doi.org/10.3390/e24040489

Chicago/Turabian Style

Zang, Huaijuan, Guoan Cheng, Zhipeng Duan, Ying Zhao, and Shu Zhan. 2022. "Automatic Search Dense Connection Module for Super-Resolution" Entropy 24, no. 4: 489. https://doi.org/10.3390/e24040489

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop