Next Article in Journal
Large-Scale Experimental Static Testing on 50-Year-Old Prestressed Concrete Bridge Girders
Next Article in Special Issue
Face Mask Detection on Photo and Real-Time Video Images Using Caffe-MobileNetV2 Transfer Learning
Previous Article in Journal
Benincasa hispida Extract Prevents Ovariectomy-Induced Osteoporosis in Female ICR Mice
Previous Article in Special Issue
DRL-FVRestore: An Adaptive Selection and Restoration Method for Finger Vein Images Based on Deep Reinforcement
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GPSR: Gradient-Prior-Based Network for Image Super-Resolution

1
College of Engineering, Huaqiao University, Quanzhou 362021, China
2
College of Fine Arts, Huaqiao University, Quanzhou 362021, China
3
School of Mathematics and Computer Science, Quanzhou Normal University, Quanzhou 362021, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(2), 833; https://doi.org/10.3390/app13020833
Submission received: 24 November 2022 / Revised: 23 December 2022 / Accepted: 3 January 2023 / Published: 7 January 2023
(This article belongs to the Special Issue AI-Based Image Processing)

Abstract

:
Recent deep learning has shown great potential in super-resolution (SR) tasks. However, most deep learning-based SR networks are optimized via pixel-level loss (i.e., L1, L2, and MSE), which forces the networks to output the average of all possible predictions, leading to blurred details. Especially in SR tasks with large scaling factors (i.e., ×4, ×8), the limitation is further aggravated. To alleviate this limitation, we propose a Gradient-Prior-based Super-Resolution network (GPSR). Specifically, a detail-preserving Gradient Guidance Strategy is proposed to fully exploit the gradient prior to guide the SR process from two aspects. On the one hand, an additional gradient branch is introduced into GPSR to provide the critical structural information. On the other hand, a compact gradient-guided loss is proposed to strengthen the constraints on the spatial structure and to prevent the blind restoration of high-frequency details. Moreover, two residual spatial attention adaptive aggregation modules are proposed and incorporated into the SR branch and the gradient branch, respectively, to fully exploit the crucial intermediate features to enhance the feature representation ability. Comprehensive experimental results demonstrate that the proposed GPSR outperforms state-of-the-art methods regarding both subjective visual quality and objective quantitative metrics in SR tasks with large scaling factors (i.e., ×4 and ×8).

1. Introduction

As one of the most crucial tasks in computer vison, Single Image Super-Resolution (SISR) aims to reconstruct a latent high-resolution (HR) image with plentiful high-frequency details from a single available low-resolution (LR) image by learning a complex nonlinear mapping. SISR has drawn much attention due to its wide range of practical applications, such as satellite imaging [1,2], medical imaging [3,4], face recognition [5,6], and video surveillance [7,8]. In recent years, deep learning has attracted increasing attention in SR tasks due to its powerful feature representation ability.
Dong et al. [9] first applied deep learning to SISR task by proposing SRCNN, which consists of three convolutional layers and which exhibits remarkable performance over traditional SR methods. Since then, numerous works have focused on deeper network structures [10], more extensive connections [11], more efficient attention mechanisms [12], and more powerful non-local operations [13,14] to enhance the feature representation of CNN-based SR models. Lim et al. [10] proposed EDSR with a very deep structure by stacking massive optimized residual blocks. Impressive performance improvement brought about by EDSR reveals that network depth plays an important role in high-quality image SR. To explore more useful SR cues from different aspects, Zhang et al. [11] proposed RDN to aggregate hierarchical features at different network depths adaptively. Further, Niu et al. [12] proposed HAN to recalibrate the hierarchical features by modeling the interdependencies among layers. In addition, Mei et al. focused on non-local operations in an SISR task and successfully proposed NLSN [13] and PA-EDSR [14], which fully utilize numerous highly similar textures within the image to synthesize the high-fidelity images.
However, most deep learning-based SR methods are optimized via pixel-level loss (i.e., L1, L2, and MSE), which guides the SR process by minimizing the pixel-to-pixel distance between the reconstructed images and the HR reference ones. Under such a constraint, the SR model tends to output the average of all possible predictions, resulting in blurred details of the reconstructed images. More seriously, the limitation is further aggravated in SR tasks with large scaling factors (i.e., ×4 and ×8). Therefore, it is important to improve the high-frequency detail reconstruction capability of the pixel-level loss-based SR models.
Considering that the gradient map of an image can reflect the spatial distribution of the high-frequency information, but also identify the local regions where the high-frequency components exist, we attempt to exploit the gradient map to guide the network to restore the appropriate edge and texture details at the suitable positions. To this end, we fully explore the potential of the gradient prior for the SR task and then propose a detail-preserving Gradient-Guided Strategy (GGS) to attain sharp edge and texture details, and to preserve the geometric invariance of the reconstructed images. Specifically, on the one hand, due to the strong correlation between the gradient information and high-frequency components, the gradient information can be transformed into the sharp edge and texture features. Thus, we introduce an additional gradient branch in the classical network with a single SR branch. The introduced gradient branch works as a feature selector that adaptively extracts structural features to provide a gradient prior for the SR branch. On the other hand, we propose a compact Gradient-Guided (GG) loss function to avoid blindly restoring high-frequency details by constraining the spatial structure of the reconstructed images. The proposed GGS facilitates the suppression of undesired geometric distortions while preventing over-smoothing or over-sharp recovery.
Furthermore, the benefit from multi-level contextual residual features, which are hierarchical with different receptive fields, can provide valuable cues for SR task from different perspectives. We propose two Residual Spatial Attention Adaptive Aggregation Modules (RS3AMs) to fully utilize the hierarchical features from the original LR images and embed them in the SR branch and the gradient branch, respectively. Specifically, on the one hand, considering that most SR methods [10,13,15] suffer from the hierarchical features being learned in a local way, which hinders the SR performance, we propose a novel Contextual Residual Fusion Structure (CRFS). Our CRFS removes over-dense connections and frequent concatenation operations from the dense block of RDN [11], and adaptively learns global hierarchical features in a holistic way, which promotes its feature representation ability. On the other hand, to enhance the feature representation of CRFS, we propose an efficient Large-Receptive-field-based Spatial Attention Module (LRSAM) to adaptively recalibrate the local spatial information of the feature map. To the best of our knowledge, large receptive fields facilitate the network to fully exploit the interdependence of local spatial information and significantly strengthen the sensitivity of the SR model toward critical spatial content. Finally, we combine LRSAM with CRFS to acquire two RS3AMs for effective feature extraction in the SR branch and the gradient branch, respectively. It is worth mentioning that GPSR achieves better SR performance with smaller parameters (8.18M) compared to most state-of-the-art SR methods, such as SAN [15] (15.9M), HAN [12] (16.07M), and NLSN [13] (46.52M). Overall, our contributions are summarized as follows.
  • We propose a Gradient-Prior-based Super-Resolution network (GPSR) for images. The experimental results show that the proposed GPSR achieves superior performance against state-of-the-art methods in terms of subjective visual results and objective evaluation metrics.
  • We propose a detail-preserving Gradient-Guided Strategy (GGS) to prompt the model to focus more attention on the most critical high-frequency components, and to suppress undesired geometric distortions as much as possible.
  • We propose two Residual Spatial Attention Adaptive Aggregation Modules (RS3AMs) and further incorporate them in the SR branch and the gradient branch, respectively, to fully explore and utilize the intermediate features to enhance feature representation capability.

2. Related Work

2.1. Gradient Prior

Extensive studies have confirmed that using common priors, which describe natural image properties, such as sparsity [16,17], spatial smoothness [18,19], non-local similarity [13,14,15,20], and gradient prior [21,22,23], can effectively strengthen the high-frequency details of the reconstructed images. Among them, the gradient prior, as one of the most critical priors, is widely used in SR tasks since it is easy to extract and can provide more advantageous information to generate sharp edges and textures. Yang et al. [22] proposed the first recurrent network model with residual learning for SR, which introduced the gradient prior between the LR images and HR images into the SR process to enhance the detail-preserving ability. To address the problem of structural distortion in GAN-based SR methods, Ma et al. [21] proposed a structure-preserving Super-Resolution (SPSR), which introduces additional gradient prior to guide the generator to restore straight and sharp edges while suppressing undesired spatial distortions as much as possible. Considering that recovering diverse low-level image elements at a single stage is not an optimal strategy due to the different characteristics of various low-level components in natural images, Wei et al. [23] proposed the Component Divide-and-Conquer (CDC) model, which produces attention masks corresponding to smooth, edge, and corner regions by utilizing gradient prior, to acquire the corresponding structural components, and then merges these three structural components as the final output. The above studies [21,22,23] have revealed that gradient prior is able to effectively improve the high-frequency details of reconstructed images. However, for SR tasks with large scaling factors (i.e., ×4 and ×8), these methods are unable to predict sharp edge and texture details, and even produce unacceptable geometric distortions due to the inability to supplement multi-scale structural information.

2.2. Contextual Residual Feature

As the network depth increases, multi-level contextual residual features, which are hierarchical with different receptive fields, can benefit SR from different perspectives. However, numerous deep-learning-based SR methods, such as VDSR [24], EDSR [10], SAN [15], PA-EDSR [14], NLSN [13], etc., constitute a deep network by simply stacking a massive quantity of residual modules, but they do not fully utilize the complementary contextual residual features, limiting the SR performance. To improve the information flow, Huang et al. [25] proposed a densely connected convolutional network (DenseNet), which uses dense connectivity to achieve a direct connection between any two layers, significantly improving the efficiency of feature and gradient information transfer, and reducing the training difficulty. However, DenseNet is unsuitable for the SR task. On the one hand, the Batch-Normalization (BN) layers severely corrupt the contrast information of the reconstructed images. On the other hand, the pooling operation may lose crucial spatial information. To address this issue, Zhang et al. [11] designed a Residual Dense Block (RDB), which densely connects abundant local features to generate more advanced fused features while removing modules that negatively impact on the SR task. Nevertheless, over-dense connections and over-frequent concatenation operations not only require a large amount of computing resources, but they also introduce redundant feature information which hinders the feature representation. To reduce the heavy computation burden, Liu et al. [26] proposed a Residual Feature Aggregation framework (RFA). Compared with RDB, RFA removes unnecessary skip connections and directly forwards the features on each local residual branch to the end of the network, leading to a better SR performance. Additionally, inspired by the feedback mechanism that permits the network to carry a signal of output to correct previous states, Li et al. [27] propose an image Super-Resolution FeedBack Network (SRFBN) to refine low-level representations with high-level information in a top-down manner to promote the SR performance. The above studies [11,25,26,27] have shown that the rational exploration and utilization of contextual residual features can provide more helpful information for the SR task. However, due to the lack of discriminative learning capability, a considerable number of redundant features may be passed to the end of the network and hinder the SR performance.

2.3. Attention-Based Networks

The visual attention mechanism is a visual information-processing mechanism unique to the human brain, which is able to acquire crucial local regions while suppressing irrelevant information by swiftly scanning the global image [28]. In the field of machine vision, the attention mechanism can be interpreted as an efficient way to manage resources by allocating more available computational resources to critical image signals. In recent years, numerous studies [12,15,29,30] have verified the effectiveness of attentional mechanisms. Considering that treating channel-wise features equally hinders the feature representation ability, Zhang et al. [29] proposed a very deep residual channel attention network (RCAN), which explores interdependencies among channels using the proposed Residual Channel Attention Module (RCAB) to adaptively reweight channel-wise features, thus prompting the network to focus on critical information. Dai et al. [15] proposed a Second-Order Attention Network (SAN), which endeavors second-order feature statistics to fully explore the dependencies of intermediate features to improve the feature representation ability further. Mei et al. [30] proposed the first Cross-Scale Non-Local (CS-NL) attention module, which combines the new CS-NL prior with local and in-scale non-local priors to reconstruct high-quality images. Considering that most attention mechanisms act only on each independent network layer and they ignore the relevance of intermediate layer features, Niu et al. [12] proposed the holistic attention network (HAN), which models the interdependencies between intermediate layers and adaptively highlights significant multi-level features. The above studies [12,15,29,30] have confirmed that the attention mechanism is able to adaptively highlight crucial features and effectively enhance the feature representation ability. However, existing spatial attention modules [31,32] only capture spatial information that are beneficial for SR tasks in extremely localized ways due to the lack of a large receptive field. Although the non-local operations [13,14,15,30,33] provide more global information to the network by capturing long-range dependencies, their high computational cost is not negligible.

3. Method

Most SR methods suffer from blindly recovering high-frequency details, resulting in their reconstructed images lacking sharp edge and texture details, along with undesired geometric distortion and artifical artifacts [34,35]. Particularly, this problem becomes more severe as the scaling factor increases. To address the problem, we propose a Gradient-Prior-based Super-Resolution network (GPSR), which is able to reconstruct sharp edge and texture details in SR tasks with large scaling factors (i.e., ×4 and ×8). In this section, we first introduce the overall architecture of our GPSR, followed by a detailed description of the detail-preserving Gradient-Guided Strategy (GGS) and the Residual Spatial Attention Adaptive Aggregation Module (RS3AM).

3.1. Network Architecture

As shown in Figure 1, our GPSR is composed of two branches, including the gradient branch and the SR branch. Among them, the gradient branch contains several gradient modules, which are mainly responsible for inferring multi-scale spatial structural feature maps to provide additional gradient priors for each decoding stage of the SR branch, thus improving the edge and texture details of the reconstructed images; while the SR branch contains several encoder–decoders, and it aims to generate the reconstructed images with high fidelity.
Given a LR image I L R as the input of our GPSR, the bicubic upsampling is first used to rescale I L R to the desired resolution, and then the shallow feature F 0 is extracted from I L R through a convolutional layer (Conv),
F 0 = f c o n v ( f u p ( I L R ) ) ,
where f u p · denotes the upsampling function with bicubic kernel and f c o n v · denotes the convolution operation.
The shallow feature map F 0 passes through N stacked encoders, yielding N shallow feature maps with different scales additionally. Let F i 1 be the input of the i-th encoder; the corresponding output F i can be obtained by:
F i = H e n ( F i 1 ) ,
where H e n · denotes the encoder function and i = 1 , 2 , . . . , N . The goal of the encoder, which contains two stacked Convs (with stride 2) and a Rectified Linear Unit (ReLU), is to perform ×2 downsampling (halving the spatial dimension and doubling the channel dimension) and shallow feature extraction operations:
H e n · = f c o n v ρ f s c o n v · ,
where f s c o n v · represents the function of the strided Conv (with stride 2), and ρ · represents Leaky ReLU.
To compensate for the information loss caused by forward propagation, we pass these shallow features F i to the decoders at each level through skip connections, where i = 0 , 1 , , N 1 .
Then, the output F N of the N-th encoder passes through N decoders, yielding N deep feature maps with different scales. Let F j + N 1 be the input of the j-th decoder; the output F j of the j-th encoder can be obtained by:
F j = H d e F j + N 1 ,
where H d e · denotes the decoder function, j = 1 , 2 , . . . , N . The goal of the decoder, which contains M stacked RS3AM-SR (Section 3.3.3) and a pixel-shuffle [36] layer followed by a Conv, is to perform ×2 upsampling (halving the channel dimension and doubling the spatial dimension) and deep feature extraction operations:
H d e · = f c o n v f s u b p i x f d e M . . . f d e 2 f d e 1 · . . . ,
where, f s u b p i x · represents the sub-pixel convolution operation, f d e a · represents the function of the a-th RS3AM-SR, and a = 1 , 2 , . . . , M . Subsequently, by concatenating the shallow feature F N j , the output F j of the j-th decoder and the spatial structure feature map F j g r a d produced by the j-th gradient module, the advanced fused feature F j + N is acquired,
F j + N = F N j , F j , F j g r a d .
Then, F j + N is treated as the input of the (j + 1)-th decoder.
Finally, the reconstructed image I S R is generated with a Conv:
I S R = f c o n v ( F 2 N ) .
where F 2 N denotes the final informative feature obtained through a series of decoders.

3.2. Detail-Preserving Gradient-Guided Strategy (GGS)

Considering that the gradient magnitude is able to effectively reflect the frequency of the image signal, in other words, the gradient magnitude is higher in the edge and texture regions corresponding to the high-frequency regions, and lower in the smooth regions corresponding to the low-frequency regions, the gradient information is able to highlight the high-frequency components beneficial for the SR task and filter out redundant low-frequency components. Consequently, a detail-preserving Gradient-Guided Strategy (GGS) is proposed to take advantage of the gradient prior from the LR images. Specifically, we design a novel gradient branch to adaptively strengthen high-frequency details and to provide gradient prior for the SR task. At the same time, we propose a compact Gradient-Guided (GG) loss that leads GPSR to infer appropriate gradient information at the suitable positions by learning the gradient space to avoid blindly recovering details.

3.2.1. Gradient Branch

In the GPSR model, the additional gradient branch is designed to achieve the spatial distribution translation from the LR gradient images to the HR gradient images. Since the gradient map of an image can be acquired by calculating the difference among neighboring pixels, we perform convolution operations with specific kernels on an LR image I L R to attain its gradient map I L R g r a d :
I L R g r a d = g I L R = I L R M x , I L R M y 2 ,
where g · , ∗ and · 2 denote the gradient feature extraction function, the convolution operation, and the 2 norm, respectively. M x and M y separately represent the fixed convolution kernels for computing the horizontal and vertical gradients:
M x = 0 0 0 1 0 1 0 0 0 , M y = 0 1 0 0 0 0 0 1 0 .
As we know, the gradient map of an image contains a wealth of high-frequency information that can be converted into sharp edge and texture details. Therefore, the gradient branch is used to restore a series of multi-scale spatial structure feature maps through learning the complex mapping relationship between the gradient maps of the LR image and the corresponding gradient maps of the HR image. All spatial structure feature maps are delivered to each decoder in the SR branch, as depicted in Equations (4)–(6), to provide multi-scale gradient priors for the SR process.
Given an LR image I L R as the input of the gradient branch, we firstly convert I L R into the gradient map fo the LR image by utilizing the gradient extraction function g · , and then we extract its shallow gradient feature F 0 g r a d with convolution operation:
F 0 g r a d = f c o n v g I L R .
To restore beneficial spatial structure feature maps, it is common practice to stack a large number of basic blocks to form a very deep neural network for SR [35]. However, such an approach severely increases the number of parameters and leads to a bloated network. Considering that the intermediate features generated on the SR branch involve abundant high-frequency information, which are significant for structure preservation, we integrate these features into the gradient branch to recover more spatial structure feature maps with fewer parameters and better reconstruction accuracy. As shown in Figure 2, the output F k 1 g r a d of the ( k 1 )-th gradient module and the intermediate features F d e , k m produced by the m-th RS3AM-SR module in the k-th decoder on the SR branch are fed into the k-th gradient module to attain the spatial structure feature map F k g r a d :
F k g r a d = f c o n v f s u b p i x H G E F k 1 g r a d , F d e , k 1 , . . . , F d e , k M ,
where H G E · denotes the function of RS3AM-G module (Section 3.3.3), and F d e , k m denotes the output of the m-th RS3AM-SR of the k-th decoder in SR branch, m = 1 , 2 , . . . , M , k = 1 , 2 , . . . , N . Then, F k g r a d is transferred to the encoder instantly, supplementing the SR process with the essential structural information, as depicted in Equations (4)–(6). The proposed gradient branch contains N gradient modules, which generate N spatial structure feature maps at different scales to provide multi-scale gradient priors for restoring more sharp edge and texture details.

3.2.2. Gradient-Guided (GG) Loss

To avoid blindly restoring high-frequency details, we propose a compact GG loss that drives the network to generate appropriate gradient information at the suitable position. The proposed GG loss includes two items that affect the gradient branch and the SR branch, respectively.
On the one hand, inspired by ProSR [37] and LapSRN [38], we impose the deep multi-scale supervision strategy on the proposed gradient branch to provide accurate spatial structure information for the SR process. Specifically, the spatial structure feature map F g r a d s generated by the s-th gradient module passes through a Conv, yielding the corresponding gradient map G s :
G s = f c o n v F g r a d s ,
where s = 1 , 2 , , N . Each G s has its own loss and corresponding label G s , which is obtained via downsampling and gradient feature extraction operations on HR images I H R :
G s = g f d o w n I H R ,
where f d o w n · represents the downsampling function with bicubic kernel. Since each G s and corresponding G s have the same shape, the gradient branch can be optimized by L S R p i x :
L g r a d p i x = s = 1 N E G s G s 1 ,
where · 1 denotes the 1 norm.
On the other hand, we impose a gradient restriction on the reconstructed images, which enables the network to intensively learn the image gradient space. Such a scheme cannot only enhance the structure preservation capacity, but also alleviate over-smoothing or over-sharpening restoration. Specifically, we restrain the reconstructed image I S R by minimizing the distance of the gradient map between the HR image I H R and the reconstructed image I S R ,
L S G p i x = E g I S R g I H R 1 .
Then, a compact GG loss L g g is formed by combining L g r a d p i x and L S G p i x :
L g g = α L g r a d p i x + β L S G p i x ,
where α and β are the parameters to balance the L g r a d p i x item and the L S G p i x item, respectively. Empirically, α and β were set to 0.1 and 0.001, respectively. Consequently, an overall loss L G P is proposed by combining pixel-wise and gradient-wise errors:
L G P = L S R p i x + L g g ,
where L S R p i x is the L 1 loss that minimizes the pixel-wise distance between the HR image I H R and its reconstructed image I S R , and can be formulated as:
L S R p i x = E I S R I H R 1 .
Ultimately, the proposed GPSR is able to effectively recover gradient information and produce the reconstructed images with sharp edges and textures by minimizing L G P .

3.3. Residual Spatial Attention Adaptive Aggregation Module

The fact that the hierarchical residual features available along the network depth are able to guide the SR process from different perspectives is ignored by most SR methods [10,11,13]. To address this issue, we propose a novel Contextual Residual Fusion Structure (CRFS) to fuse complementary contextual residual features. Further, to enhance the feature learning capability of the CRFS, we propose an efficient Large-Receptive-field-based Spatial Attention Module (LRSAM) to adaptively recalibrate the spatial feature responses toward the most crucial of the input. The large receptive field is required for the attention block to work well for exploiting the interdependence of local information. Finally, we propose two residual spatial attention adaptive aggregation modules (i.e., RS3AM-G and RS3AM-SR), consisting of CRFS and a set of LRSAMs, for powerful feature extraction in the gradient branch and the SR branch, respectively.

3.3.1. Contextual Residual Fusion Structure (CRFS)

To take full advantage of the complementary contextual residual features, we propose a novel Contextual Residual Fusion Structure (CRFS). Different to the conventional single-path feed-forward framework, we redesign the arrangement of the stacked residual blocks [10], and then we add three residual blocks on the residual branch of CRFS to adaptively preserve the hierarchical features in a global way, as shown in Figure 3. Specifically, all contextual residual features are passed to a gating unit, which concatenates these multi-level features and adaptively learns the weights of different residual features through convolution operations. Let R i , d denote the i-th residual feature of the d-th CRFS, where i = 1 , 2 , 3 , 4 , then the corresponding fused feature F c o n , d is computed by:
F c o n , d = H g a t e R 1 , d , R 2 , d , R 3 , d , R 4 , d
where H g a t e · denotes the function of the gate unit. To improve the information flow and to compensate for the information loss during forward propagation, global residual learning is then employed to produce the output F d of the d-th CRFS:
F d = F d 1 + F c o n , d
where F d 1 denotes the output of the ( d 1 )-th CRFS.

3.3.2. Large-Receptive-Field-Based Spatial Attention Module

To promote the feature learning capability of CRFS, introducing a spatial attention mechanism is a reasonable solution. However, some spatial attention modules [32,33] are unable to effectively capture spatial location information that is helpful for SR task, due to the lack of large receptive fields. Non-local operations [13,14,15,30,33] serves as a promising solution to implement a spatial attention mechanism that explores correlations between any positions in a global way by capturing long-range dependencies. However, the large computational overhead is an insurmountable drawback. Considering that large receptive fields are able to fully exploit the interdependencies of local positional information, we propose an efficient Large-Receptive-field-based Spatial Attention Module (LRSAM) to enhance the sensitivity of the SR model to critical spatial content. Compared to non-local operations, LRSAM has a lower computational complexity and can be easily embedded in different network positions.
LRSAM consists of two sub-modules, including the squeeze module and the excitation module, as shown in Figure 4. The squeeze module starts with a Conv to reduce the channel number of the input feature Y 0 , then yielding a large receptive field by a stride Conv (with stride 3) and the max-pooling operation with a large window (e.g., 7 × 7) for enhancing the sensitivity of critical spatial content. To generate the spatial attention map, the squeeze module firstly applies a set of Convs followed by ReLU on the output Y 1 of the squeeze module. In the excitation module, Y 1 is passed through an upsampling layer of pixel-shuffle [36] and sigmoid activation to produce a spatial attention map Y m a s k , and then Y m a s k is used to rescale Y 0 for highlighting beneficial feature information and to obtain Y 0 * .

3.3.3. RS3AM-SR and RS3AM-G

To further improve the performance of CRFS, LRSAM is incorporated into the CRFS to construct two Residual Spatial Attention Adaptive Aggregation Modules (i.e., RS3AM-SR, RS3AM-G), which are embedded in the SR branch and the gradient branch, respectively, for efficient feature extraction. RS3AM-SR and RS3AM-G differ in their structural designs, depending on the characteristics of the different branches.
The architecture of RS3AM-SR that acts on the SR branch is depicted in Figure 5a. We embed a series of LRSAMs at the tails of residual branches at each level of CRFS. These LRSAMs are able to highlight spatial features that are useful for the SR task, and these highlighted features are passed through the gate unit to produce more beneficial features. Specifically, let R c , d s r denotes the c-th residual feature of the d-th RS3AM-SR, which is enhanced by LRSAMs, where c = 1 , 2 , . . . , M , then the corresponding fused feature F c o n , d s r is computed by:
F c o n , d s r = H g a t e R 1 , d , R 2 , d , . . . , R M , d .
To compensate for the information loss during forward propagation and to improve the information flow, global residual learning is then employed to obtain output F d s r of the d-th RS3AM-SR:
F d s r = F c o n , d s r + x ,
where x denotes the input of the d-th RS3AM-SR.
The architecture of RS3AM-G that acts on the gradient branch is depicted in Figure 5b. As the intermediate features of the SR branch carry a wealth of high-frequency information, RS3AM-G utilizes them as an essential complement to inferring sharp spatial structure features. Specifically, the feature map that is concatenated by the output F d e , k m of the m-th RS3AM-SR of the k-th decoder and the local residual feature F m , k l o c a l on m-th residual branch of the k-th RS3AM-G passes through an LRSAM and a Conv, yielding a fused feature R m , k g r a d :
R m , k g r a d = f c o n v f L R S A F d e , k m , F m , k l o c a l ,
where f L R S A denotes the function of the LRSAM, m = 1 , 2 , . . . , M , k = 1 , 2 , . . . , N . It should be noted that the function of Conv here is to halve the channel dimension.
Subsequently, local feature fusion [39] is applied to adaptively fuse all intermediate features R m , k g r a d in the current RS3AM-G, and then local residual learning is introduced in RS3AM-G to introduce the input x of the k-th RS3AM-G to further improve the information flow. Thus, the output F k G of the k-th RS3AM-G can be obtained by:
F k G = H g a t e R 1 , k g r a d , R 2 , k g r a d , . . . , R M , k g r a d + x .
In contrast to dense block [25] and its variants [11], our RS3AM removes modules that have negative impacts on the SR performance, such as batch normalization (BN) layers, while simplifying excessive dense connections and frequent concatenating operations, and reinforces the spatial distribution of residual features at all levels by using LRSAM. Hence, the proposed RS3AMs are sufficiently efficient and lightweight. Finally, we used both RS3AM-SR and RS3AM-G to construct GPSR. Compared to most state-of-the-art SR methods, such as SAN (15.9M), HAN (16.07M), PA-EDSR (45.53M), NLSN (46.52M), etc., GPSR achieves a better super-resolution with fewer parameters (8.18M).

4. Experiments

In this section, we first present the experimental settings regarding datasets, degradation models, quantitative metrics, and training settings, then we visually validate the effectiveness of the Gradient-Guided Strategy (GGS), and comprehensively analyze the contributions of the proposed three modules, including Large-Receptive-field-based Spatial Attention Module (LRSAM), Contextual Residual Fusion Structure (CRFS), and GGS. Finally, we compare the proposed Gradient-Prior-based Super-Resolution network (GPSR) with the state-of-the-art methods in terms of both quantitative metrics and visual quality.

4.1. Settings

4.1.1. Datasets and Degradation Models

Following [29,40], DIV2K [41] were selected as training sets to train the proposed GPSR, while five commonly benchmark datasets were used for testing, including set5 [42], set14 [43], BSD100 [44], Urban100 [45], and Manga109. Table 1 reports the characteristics of each dataset. Our experiments are separately conducted with bicubic (BI) and blur-downscale (BD) degradation models. In each training batch, 8 LR color patches with the size of 48 × 48 are extracted as the input of our GPSR.

4.1.2. Quantitative Metrics

To evaluate the quality of the reconstructed images, we first convert the obtained reconstructed images into YCbCr space and then calculate the corresponding Peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) metrics on the Y channel. This is because it is more efficient to calculate the metrics on the Y channel of the converted YCbCr space than on the three channels of the original RGB color space. It is worth noting that the high metric means a better super-resolution performance.

4.1.3. Training Settings

For training, we apply the Adam optimizer [46] with β 1 = 0.9 , β 2 = 0.999 , ε = 10 8 . This is because the Adam optimizer has the advantages of fewer hyperparameters, efficient computation, and fast convergence. The learning rate is initialized to 10 4 and decreased to 10 7 with a cosine annealing out of 10 6 iterations in total. Furthermore, in an experiment with scaling factor S (i.e., ×4, ×8), the numbers of encoders, decoders, and gradient modules are set to N = log 2 S ; the number of RS3AM-SR and the number of residual branch of RS3AM-G are set to M = 4 . Our GPSR has been executed on the PyTorch framework and on an Nvidia GeForce RTX 3090 24GB GPU.

4.2. Ablation Experiments

We analyze the effectiveness of the proposed modules, including GGS, LRSAM, and CRFS, via ablation experiments. The baseline model is built upon the proposed GPSR by removing the GB and replacing the RS3AM-SR modules, which consist of CRFS and LRSAM, of the SR branch with the plain residual modules [10].
Analysis on the gradient branch: To verify the effectiveness of the GB, we visualize the output of the GB. Figure 6 illustrates the visual results of the GB. As can be seen from Figure 6a,d, the gradient maps of the HR images, which not only contain the sharp edge, but also effectively reflect the spatial structure relationship, are obtained by performing gradient feature extraction on the original HR images. This motivates us to design a detail-preserving GB to generates the gradient maps of the reconstructed images containing sharp geometric structures, and provides the essential gradient guidance for the SR process. Specifically, through the introduced GB, the gradient maps of the LR images (shown in Figure 6b,e) were converted to the gradient maps of the reconstructed images (shown in Figure 6c,f). It can be observed that the gradient maps of the LR images are blurred and smooth in general, while the spatial structure details contained in the gradient maps of the reconstructed images (shown in Figure 6c,f) are significantly sharper and highly similar to the gradient maps of the HR images (shown in Figure 6a,d). This experimental result shows that the proposed GB possesses the ability to restore high-fidelity edge and texture information. This is because the proposed GB is able to generate the gradient maps at different scales and simultaneously transmit them as gradient priors for each upsampling stage on the SR branch to complement the multi-scale spatial structure features.
Analysis on gradient-guided loss: We analyze the effectiveness of gradient-guided (GG) loss by removing L g r a d p i x or L S G p i x from the baseline model with the GGS, and report the results in Table 2. Here, L g r a d p i x is used to supervise the outputs of the GB to ensure the accuracy of the multi-scale gradient maps. Additionally, L S G p i x is used to supervise the gradient map of the final SR image. When L g r a d p i x is removed, the GB of the corresponding model cannot be optimized, and it outputs unreliable multi-scale gradient maps, resulting in SR performance degrading from 32.61 dB to 32.31 dB. When L S G p i x is removed, the SR performance of the corresponding model is reduced from 32.61 dB to 32.54 dB. It is because L g r a d p i x can effectively avoid the SR model from blindly restoring high-frequency information. From the above analysis, optimizing L g r a d p i x and L S G p i x jointly is essential. Consequently, the proposed gradient-guided loss can guide the network to reconstruct accurate high-frequency details by taking full advantage of L g r a d p i x and L S G p i x .
Analysis on gradient guidance strategy: We visually compared the baseline with and without the GGS, including the GB and the GG loss. Specifically, the LR images (shown in Figure 7a,d) as the network input, the reconstructed images, and the corresponding absolute gradient intensity maps (shown in Figure 7b,e) obtained using the model without GGS, and the reconstructed images and the corresponding absolute gradient intensity maps (shown in Figure 7c,f) obtained using the model with GGS, are depicted in Figure 7. Notably, the absolute gradient intensity maps directly reflect the quality of edge and texture details of the reconstructed images by using the brightness of color. The brighter the color is, the sharper the structure details are. From Figure 7b,c, it can be seen that the sharpness of the whiskers in the reconstructed image is significantly improved by adopting GGS. Additionally, compared with Figure 7e, the wire on the window in Figure 7f appears sharper, which is more evident in the absolute gradient intensity map in Figure 7f. Additionally, the PSNR of the reconstructed images (i.e., Figure 7c,f) obtained by the model with GGS is significantly improved by 0.09 dB and 0.38 dB, respectively, compared to the baseline. It indicates that the proposed GGS not only strengthens the edge and texture details, but also alleviates the geometric distortion well.
Analysis on effects of CRFS, CRFS, and GGS: We conduct ablation experiments to analyze the effectiveness of CRFS, CRFS, GGS and their combinations, and report the results in Table 3. Compared with the baseline, the models with only CRFS or LRSAM achieve better results by up to 0.06 dB and 0.06 dB in terms of the PSNR metric, respectively. These results demonstrate that fusing the contextual residual feature can provide more advanced features for SR task, and spatial attention mechanisms with large-sized receptive fields can effectively explore valuable spatial content. At the same time, compared with the baseline, the PSNR obtained by the model with only GGS improves by 0.15 dB. It can be seen that our GGS possesses the highest contribution to performance improvement. This is because GGS provides the beneficial gradient prior for the SR process, which facilitates the guidance of the GPSR model to restore sharp high-frequency details. Furthermore, the improvement of the model with both CRFS and LRSAM is significant, with a PSNR improvement of 0.17 dB over the one with only CRFS, which indicates that LRSAM is able to effectively promote the feature representation ability of CRFS. Finally, we obtained the highest gain (0.28 dB) by the model with LRSAM, CRFS, and GGS simultaneously, which is the GPSR model.

4.3. Results with Bicubic (BI) Degradation

We compare the proposed GPSR with the nine state-of-the-art methods, including LapSRN [38], HAN [12], DBPN [34], RDN [11], EDSR [10], SAN [15], PA-EDSR [14], DRN [40], and NLSN [13] in terms of quantitative results and visual quality. Following [10,29], we also propose a self-ensemble model and denote it as GPSR+.

4.3.1. Quantitative Metrics

For quantitative comparison, we compare the PSNR and SSIM metrics of different methods for the SR task with scaling factors ×4 and ×8. Table 4 lists the comparison of ×4 and ×8 quantitative results on the five commonly benchmark datasets, including Set5, Set14, BSD100, Urban100, and Manga109, where the optimal and suboptimal metrics are highlighted in red and blue, respectively, and where ‘–’ means that the result is not available. Params and FLOPs denote the total number of parameters and floating-point operations, respectively. Noted that each the efficiency proxy FLOPs is measured under the setting of upscaling reconstructed images to the size of 1280 × 720 on the scaling factors ×4 and ×8. It should be noted that we derive the results of the methods used for comparison from their pre-trained models, released code, or their original paper. It is worth noting that some comparison methods, including HAN, RDN, EDSR, PA-EDSR, and NLSN, do not offer experimental results for ×8 SR. Therefore, we retrained these models on the ×8 scaling factor with their official source codes. In addition, for a fair comparison, we retrained DRN on the DIV2K dataset, as it originally used the DF2K dataset as the training set.
For convenience to quantitative comparisons, we take Set5 (×4) as an example. From Table 4, we observe that SR models #1 with deeper network structures and more extensive connections, including DBPN, RDN, and EDSR, obtain approximately 0.9 dB of PSNR gains compared to LapSRN. Further, SR models #2 with attention mechanisms, including HAN, SAN, PA-EDSR, and NLSN, can adaptively emphasize useful feature information for more accurate image reconstruction. The PSNR gains of models #2 over #1 range from 0.12 dB to 0.19 dB. However, the above models ignore the potential of the gradient prior for image SR. In contrast, the proposed GPSR makes full use of the gradient map to restore sharp high-frequency details, and obtains the highest quantitative metrics. It is worth noting that the proposed GPSR achieves a performance advantage of 0.09 dB with fewer parameters (18% parameters of PA-EDSR) and FLOPs (34% Flops of PA-EDSR) compared to PA-EDSR, which obtains sub-optimal SR results.This indicates that the proposed GPSR takes into account the computational efficiency while pursuing performance.
From Table 4, we can see that GPSR shows superior performance for both ×4 and ×8 SR tasks. GPSR+ achieves the optimal results on all test datasets. Compared with the nine state-of-the-art methods, the PSNR and SSIM metrics obtained by our GPSR are optimal in most test datasets. Specifically, our GPSR performs optimally on all datasets for both ×4 and ×8 SR tasks, except for the suboptimal SSIM metric on BSD100 for the ×4 SR task. Consequently, Table 4 provides adequate evidence that the proposed GPSR achieves the optimal super-resolution performance in the SR tasks with large scaling factors. Furthermore, Table 4 also presents the number of parameters and FLOPs for different SR methods. Overall, the proposed GPSR achieves a better tradeoff between computational efficiency and the super-resolution performance.
We also present an illustrated comparison in terms of the PSNR metric and the number of parameters for both the ×4 and ×8 SR tasks in Figure 8. As can be seen from Figure 8, compared with other SR models used for comparison, the proposed GPSR achieves the highest PSNR with a small number of parameters on the Set5 dataset for both the ×4 and ×8 SR tasks, indicating its optimal super-resolution performance.

4.3.2. Visual Quality

Figure 9 illustrates the visual comparisons of ×4 and ×8 SR tasks, respectively, which offer the reconstructed results obtained using different methods in the same patches, and the corresponding original HR patches are given as references. From Figure 9, we can see that most of the methods used for comparison are weak in reconstructing sharp edge and texture details, and even produce severe geometric distortions and artificial artifacts. Compared with these methods, our GPSR is able to reconstruct a clearer image with sharp edge and texture details. Remarkably, GPSR suppresses the undesired geometric distortions well, and this advantage becomes increasingly important as the scaling factor increases in SR tasks.
In Figure 9a, we show visual comparisons of the ×4 SR task. For the image “barbara” from the Set14 dataset, we observe that most of the compared methods, such as LapSRN, EDSR, RDN, and DRN fail to recover the edge details of the book effectively, making their reconstructed images appear blurred and smooth. Compared with the previous methods, SAN and NLSN show better detail recovery ability, but their reconstructed images are still unsatisfactory. Compared with the compared methods, the proposed GPSR shows better structure preservation and detail recovery performance, which is reflected in sharper and more complete edge details. From the image “253027” in the BSD100 dataset, the zebra stripes produced by most of the compared methods are blurred and confused, and there exist different degrees of information loss in the local regions. However, DRN, SAN, and NLSN retain more texture than LapSRN. However, they are still inferior to our GPSR. This is due to the fact that benefits from our GGS, the proposed GPSR, are able to effectively local gradient information to reconstruct high-frequency details while avoiding the problem of blindly restoring details that exist in most comparative models.
Figure 9b presents visual comparisons of the ×8 SR task. For the image “butterfly” from Set5, we observe that most of the compared methods fail to restore the spots and suffer from blurred artifacts. In contrast, our GPSR is able to alleviate the blurred artifacts better, and reconstruct sharp edge and texture details. For the image “img040” from Urban100, the reconstructed images produced by most comparative methods have significant geometric distortions, such as significant errors in the extension directions of the edges. Although HAN, RDN, and DBPN are able to suppress geometric distortions to a certain extent effectively, this deficiency still exists in local regions caused by blind detail restoration. The optimal reconstructed image is generated by the proposed GPSR, which again proves the effectiveness of the proposed GGS.

4.4. Results with Blur-Downscale (BD) Degradation

Following [29], we also test the effectiveness of the proposed GPSR on the LR images with the blur-down (BD) degradation model. We selected the five state-of-the-art methods, including RCAN [29], DRN [40], HAN [12], PA-EDSR [14], and NLSN [11], to compare with our GPSR in objective quantitative metrics. Table 5 depicts the results of different SR models for the ×8 SR task. It can be seen from Table 5 that our GPSR also consistently outperforms other compared methods, even without self-ensemble in the SR task with large scaling factor. Specifically, the PSNR gain of our GPSR over NLSN is up to 1.05 dB and 1.25 dB on the Urban100 and Manga109 datasets, respectively.

5. Conclusions

In this paper, we fully explore the potential of the gradient prior for SR task. Specifically, to restore sharp high-frequency details, we introduce an additional gradient branch in the classical SR network to provide the beneficial structural features for each upsampling stage of the SR process. Meanwhile, we propose a compact Gradient-Guided (GG) loss to strengthen the constraints on the spatial structure of the reconstructed images, so as to guide the model to restore the appropriate gradient information at the suitable position and to avoid blindly restoring high-frequency details. Additionally, we propose a novel Contextual Residual Fusion Structure (CRFS), which is capable of fully fusing complementary contextual residual features to produce more advanced features; further, to promote the feature representation ability of CRFS as much as possible, we propose an efficient Large-Receptive-field-based Spatial Attention Module (LRSAM) to highlight the critical residual features; finally, we incorporate LRSAM into CRFS and further propose two Residual Spatial Attention Adaptive Aggregation Modules (RS3AMs) used for image feature extraction and gradient feature extraction, respectively. Extensive experiments indicate that the superiority of our Gradient-Prior-based Super-Resolution network (GPSR) for images over state-of-the-art methods for SR tasks with large scaling factors (×4 and ×8). However, there are two limitations to our method, as follows: (1) Although our GPSR achieves better results with fewer parameters and FLOPs than most existing SR methods, its computational efficiency still needs to be further improved to satisfy the requirements of real applications, such as the deployment of the technique in embedded devices. (2) Like most existing SR methods, the proposed GPSR only considers known degradation models (e.g., BI and BD degradation kernels), and it is still struggling with SR tasks with unknown degradation kernels. To address these two limitations, we will focus on compressing the model to improve computational efficiency while retaining SR performance in our future work. Moreover, we will explore the introduction of the adversarial generative network into the SR task to estimate the unknown degradation kernels, which facilitates the real-world image SR.

Author Contributions

X.Z.: Conceptualization, Methodology, and Writing. D.H.: Supervision, Review, and Editing. X.L.: Software and Visualization. D.C.: Original draft preparation. D.Z.: Editing and Review. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 61901183, in part by the National Key R&D Program of China under the grant 2021YFE0205400, in part by the Collaborative Innovation Platform Project of Fujian Province under Grant 2021FX03, in part by the Natural Science Foundation of Fujian Provincial Science and Technology Department under Grant 2021H6037, in part by the Natural Science Foundation of Fujian Province under Grant 2019J01010561, in part by the Fundamental Research Funds for the Central Universities Grant ZQN-921, in part by the Foundation of Fujian Education Department under Grant JAT170053, and in part by the Key Project of Quanzhou Science and Technology Plan under Grant 2021C008R.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

https://pan.baidu.com/s/1hXuPoIHf8A4wm1ZzM7rBfQ?pwd=Z983 (accessed on 23 November 2022), password: Z983.

Acknowledgments

In this section you can acknowledge any support given which is not covered by the author contribution or funding sections. This may include administrative and technical support, or donations in kind (e.g., materials used for experiments).

Conflicts of Interest

No potential conflict of interests were reported by the authors. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Zou, F.; Xiao, W.; Ji, W.; He, K.; Yang, Z.; Song, J.; Zhou, H.; Li, K. Arbitrary-oriented object detection via dense feature fusion and attention model for remote sensing super-resolution image. Neural Comput. Appl. 2020, 32, 14549–14562. [Google Scholar] [CrossRef]
  2. Lei, S.; Shi, Z.; Zou, Z. Super-resolution for remote sensing images via local–global combined network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1243–1247. [Google Scholar] [CrossRef]
  3. Peled, S.; Yeshurun, Y. Superresolution in MRI: Application to human white matter fiber tract visualization by diffusion tensor imaging. Magn. Reson. Med. 2001, 45, 29–35. [Google Scholar] [CrossRef] [PubMed]
  4. Thornton, M.W.; Atkinson, P.M.; Holland, D. Sub-pixel mapping of rural land cover objects from fine spatial resolution satellite sensor imagery using super-resolution pixel-swapping. Int. J. Remote Sens. 2006, 27, 473–491. [Google Scholar] [CrossRef]
  5. Chen, Y.; Tai, Y.; Liu, X.; Shen, C.; Yang, J. Fsrnet: End-to-end learning face super-resolution with facial priors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2492–2501. [Google Scholar]
  6. Yang, T.; Ren, P.; Xie, X.; Zhang, L. GAN prior embedded network for blind face restoration in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 672–681. [Google Scholar]
  7. Zou, W.W.; Yuen, P.C. Very low resolution face recognition problem. IEEE Trans. Image Process. 2011, 21, 327–340. [Google Scholar] [CrossRef] [PubMed]
  8. Zhang, L.; Zhang, H.; Shen, H.; Li, P. A super-resolution reconstruction algorithm for surveillance images. Signal Process. 2010, 90, 848–859. [Google Scholar] [CrossRef]
  9. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar]
  10. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  11. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2472–2481. [Google Scholar]
  12. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 191–207. [Google Scholar]
  13. Mei, Y.; Fan, Y.; Zhou, Y. Image super-resolution with non-local sparse attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3517–3526. [Google Scholar]
  14. Mei, Y.; Fan, Y.; Zhang, Y.; Yu, J.; Zhou, Y.; Liu, D.; Fu, Y.; Huang, T.S.; Shi, H. Pyramid attention networks for image restoration. arXiv 2020, arXiv:2004.13824. [Google Scholar]
  15. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.T.; Zhang, L. Second-order attention network for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11065–11074. [Google Scholar]
  16. Daubechies, I.; Defrise, M.; De Mol, C. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Commun. Pure Appl. Math. 2004, 57, 1413–1457. [Google Scholar] [CrossRef] [Green Version]
  17. Tropp, J.A.; Wright, S.J. Computational methods for sparse solution of linear inverse problems. Proc. IEEE 2010, 98, 948–958. [Google Scholar] [CrossRef] [Green Version]
  18. Marquina, A.; Osher, S.J. Image super-resolution by TV-regularization and Bregman iteration. J. Sci. Comput. 2008, 37, 367–382. [Google Scholar] [CrossRef]
  19. Aly, H.A.; Dubois, E. Image up-sampling using total-variation regularization with a new observation model. IEEE Trans. Image Process. 2005, 14, 1647–1659. [Google Scholar] [CrossRef] [PubMed]
  20. Peyré, G.; Bougleux, S.; Cohen, L. Non-local regularization of inverse problems. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; pp. 57–68. [Google Scholar]
  21. Ma, C.; Rao, Y.; Cheng, Y.; Chen, C.; Lu, J.; Zhou, J. Structure-preserving super resolution with gradient guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 7769–7778. [Google Scholar]
  22. Yang, W.; Feng, J.; Yang, J.; Zhao, F.; Liu, J.; Guo, Z.; Yan, S. Deep edge guided recurrent residual learning for image super-resolution. IEEE Trans. Image Process. 2017, 26, 5895–5907. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Wei, P.; Xie, Z.; Lu, H.; Zhan, Z.; Ye, Q.; Zuo, W.; Lin, L. Component divide-and-conquer for real-world image super-resolution. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 101–117. [Google Scholar]
  24. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  25. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  26. Liu, J.; Zhang, W.; Tang, Y.; Tang, J.; Wu, G. Residual feature aggregation network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2359–2368. [Google Scholar]
  27. Li, Z.; Yang, J.; Liu, Z.; Yang, X.; Jeon, G.; Wu, W. Feedback network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3867–3876. [Google Scholar]
  28. Itti, L.; Koch, C.; Niebur, E. A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1254–1259. [Google Scholar] [CrossRef] [Green Version]
  29. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  30. Mei, Y.; Fan, Y.; Zhou, Y.; Huang, L.; Huang, T.S.; Shi, H. Image super-resolution with cross-scale non-local attention and exhaustive self-exemplars mining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5690–5699. [Google Scholar]
  31. Hu, Y.; Li, J.; Huang, Y.; Gao, X. Channel-wise and spatial feature modulation network for single image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 3911–3927. [Google Scholar] [CrossRef]
  32. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  33. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7794–7803. [Google Scholar]
  34. Haris, M.; Shakhnarovich, G.; Ukita, N. Deep back-projection networks for super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1664–1673. [Google Scholar]
  35. Liu, Y.; Wang, Y.; Li, N.; Cheng, X.; Zhang, Y.; Huang, Y.; Lu, G. An attention-based approach for single image super resolution. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2777–2784. [Google Scholar]
  36. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  37. Wang, Y.; Perazzi, F.; McWilliams, B.; Sorkine-Hornung, A.; Sorkine-Hornung, O.; Schroers, C. A fully progressive approach to single-image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 864–873. [Google Scholar]
  38. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep laplacian pyramid networks for fast and accurate super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 624–632. [Google Scholar]
  39. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  40. Guo, Y.; Chen, J.; Wang, J.; Chen, Q.; Cao, J.; Deng, Z.; Xu, Y.; Tan, M. Closed-loop matters: Dual regression networks for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5407–5416. [Google Scholar]
  41. Timofte, R.; Agustsson, E.; Van Gool, L.; Yang, M.H.; Zhang, L. Ntire 2017 challenge on single image super-resolution: Methods and results. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 114–125. [Google Scholar]
  42. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the British Machine Vision Conference; BMVA Press: Guildford, UK, 2012; pp. 135.1–135.10. [Google Scholar]
  43. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the International Conference on Curves and Surfaces, Avignon, France, 24–30 June 2010; pp. 711–730. [Google Scholar]
  44. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision, ICCV 2001, Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  45. Huang, J.B.; Singh, A. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  46. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Network architecture of our Gradient-Prior-based Super-Resolution network (GPSR). The network consists of two branches, the gradient branch (bottom right), which is used to generate the multi-scale spatial structure feature maps, and the SR branch (upper left), which is used to generate the final reconstructed image. Notably, the intermediate features generated on the SR branch are integrated into the gradient branch to provide additional beneficial structural features. In addition, the multi-scale spatial structure feature maps generated on the gradient branch are integrated into the SR branch to strengthen the edge and texture details.
Figure 1. Network architecture of our Gradient-Prior-based Super-Resolution network (GPSR). The network consists of two branches, the gradient branch (bottom right), which is used to generate the multi-scale spatial structure feature maps, and the SR branch (upper left), which is used to generate the final reconstructed image. Notably, the intermediate features generated on the SR branch are integrated into the gradient branch to provide additional beneficial structural features. In addition, the multi-scale spatial structure feature maps generated on the gradient branch are integrated into the SR branch to strengthen the edge and texture details.
Applsci 13 00833 g001
Figure 2. Architecture of the k-th gradient module of the gradient branch.
Figure 2. Architecture of the k-th gradient module of the gradient branch.
Applsci 13 00833 g002
Figure 3. Architecture of the Contextual Residual Fusion Structure (CRFS).
Figure 3. Architecture of the Contextual Residual Fusion Structure (CRFS).
Applsci 13 00833 g003
Figure 4. Architecture of the Large-Receptive-field-based Spatial Attention Module (LRSAM).
Figure 4. Architecture of the Large-Receptive-field-based Spatial Attention Module (LRSAM).
Applsci 13 00833 g004
Figure 5. Two kinds of residual space attention adaptive aggregation module architectures (i.e., RS3AM-SR and RS3AM-G). (a) Architecture of the d-th RS3AM-SR that acts on the SR branch. (b) Architecture of the k-th RS3AM-G that acts on the gradient branch. For convenience, we only show the architectures of these two models separately when M = 4 .
Figure 5. Two kinds of residual space attention adaptive aggregation module architectures (i.e., RS3AM-SR and RS3AM-G). (a) Architecture of the d-th RS3AM-SR that acts on the SR branch. (b) Architecture of the k-th RS3AM-G that acts on the gradient branch. For convenience, we only show the architectures of these two models separately when M = 4 .
Applsci 13 00833 g005
Figure 6. Results of gradient branch. Among them, (a,d) are the gradient maps of the reference HR images, (b,e) are the gradient maps of the corresponding LR images, and (c,f) are the gradient maps of the reconstructed images generated by the gradient branch.
Figure 6. Results of gradient branch. Among them, (a,d) are the gradient maps of the reference HR images, (b,e) are the gradient maps of the corresponding LR images, and (c,f) are the gradient maps of the reconstructed images generated by the gradient branch.
Applsci 13 00833 g006
Figure 7. Visual comparison of the baseline with and without the Gradient-Guided Strategy (GGS). We present the LR images, the reconstructed images, and the absolute gradient intensity maps corresponding to the reconstructed images. Among them, (a,d) are input LR images, (b,e) are reconstructed results of baseline without GGS, and (c,f) are reconstructed results of baseline with GGS. The proposed GGS is able to effectively preserve the structure and acquire sharper high-frequency details.
Figure 7. Visual comparison of the baseline with and without the Gradient-Guided Strategy (GGS). We present the LR images, the reconstructed images, and the absolute gradient intensity maps corresponding to the reconstructed images. Among them, (a,d) are input LR images, (b,e) are reconstructed results of baseline without GGS, and (c,f) are reconstructed results of baseline with GGS. The proposed GGS is able to effectively preserve the structure and acquire sharper high-frequency details.
Applsci 13 00833 g007
Figure 8. Comparisons of the PSNR metric and the number of parameters among different SR models on the Set5 dataset.
Figure 8. Comparisons of the PSNR metric and the number of parameters among different SR models on the Set5 dataset.
Applsci 13 00833 g008
Figure 9. Visual comparisons for SR tasks. (a) Visual comparison for ×4 SR task with BI model on Set14 and BSD100 datasets. (b) Visual comparison for ×8 SR task with BI model on Set14 and Urban100 datasets.
Figure 9. Visual comparisons for SR tasks. (a) Visual comparison for ×4 SR task with BI model on Set14 and BSD100 datasets. (b) Visual comparison for ×8 SR task with BI model on Set14 and Urban100 datasets.
Applsci 13 00833 g009
Table 1. Characteristics of the dataset used for experiments.
Table 1. Characteristics of the dataset used for experiments.
DatasetNumber of ImagesSizeMain ContentFunction
DIV2K [41]9005.8 GBHumans, animals, and landscapesTraining
Set5 [42]51.30 MBHumans and animalsTesting
Set14 [43]147.44 MBHumans and animalsTesting
BSD100 [44]10038.87 MBLandscapes and animalsTesting
Urban100 [45]100194 MBBuildingsTesting
Manga109109219 MBMangaTesting
Table 2. Ablation experiments on gradient-guided loss, including L g r a d p i x and L S G p i x . The average PSNR values are evaluated on Set5 dataset with scaling factor ×4.
Table 2. Ablation experiments on gradient-guided loss, including L g r a d p i x and L S G p i x . The average PSNR values are evaluated on Set5 dataset with scaling factor ×4.
L g r a d p i x
L S G p i x
PSNR32.5432.3132.61
Table 3. Investigations of LRSAM, CRFS, and GGS. We observe that the proposed GPSR, that is, the model that uses LRSAM, CRFS, and GGS simultaneously, achieves the best super-resolution performance on Set5 dataset with scaling factor ×4.
Table 3. Investigations of LRSAM, CRFS, and GGS. We observe that the proposed GPSR, that is, the model that uses LRSAM, CRFS, and GGS simultaneously, achieves the best super-resolution performance on Set5 dataset with scaling factor ×4.
LRSAM
CRFS
GGS
PSNR32.4632.5232.5232.6132.6332.7032.6732.74
Table 4. Quantitative metrics obtained by various models for SR task with BI degradation model. The optimal and suboptimal metrics are highlighted in red and blue.
Table 4. Quantitative metrics obtained by various models for SR task with BI degradation model. The optimal and suboptimal metrics are highlighted in red and blue.
MethodParamsFLOPsScaleSet5Set14BSD100Urban100Manga109
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
LapSRN [38]0.9M- 31.54/0.885028.19/0.772027.32/0.727025.21/0.756029.09/0.8900
HAN [12]16.07M944G 32.64/0.900228.90/0.789027.80/0.744226.85/0.809431.42/0.9177
DBPN [34]10.4M1106G 32.47/0.898328.75/0.785927.67/0.739626.38/0.794730.90/0.9134
RDN [11]22.42M1309G 32.47/0.899028.81/0.787127.72/0.741926.61/0.802831.00/0.9151
EDSR [10]43.1M2894G 32.46/0.896828.80/0.787627.71/0.742026.64/0.803331.03/0.9149
SAN [15]15.9M936G432.64/0.900328.92/0.788827.79/0.743626.79/0.806831.18/0.9169
PA-EDSR [14]45.53M3214G 32.65/0.900628.87/0.789127.76/0.744527.01/0.814031.29/0.9194
DRN [40]4.8M685G 32.61/0.897428.89/0.787627.74/0.738926.71/0.803831.33/0.9150
NLSN [13]46.52M2955G 32.59/0.900028.87/0.789127.78/0.744426.96/0.810931.27/0.9184
GPSR8.18M1098G 32.74/0.901028.98/0.790127.81/0.744126.96/0.811631.50/0.9200
GPSR+8.18M- 32.82/0.902229.06/0.791327.86/0.745127.14/0.815031.80/0.9223
LapSRN [38]1.3M- 26.15/0.738024.35/0.620024.54/0.586021.81/0.581023.39/0.7350
HAN [12]16.22M271G 27.33/0.788425.24/0.651024.98/0.605922.98/0.643725.20/0.8011
DBPN [34]23.2M236G 27.21/0.784425.13/0.648924.88/0.601022.72/0.631525.50/0.7984
RDN [11]22.42M327G 26.91/0.771724.92/0.640524.85/0.599322.52/0.621124.60/0.7768
EDSR [10]45.5M1271G 27.22/0.784025.14/0.647624.88/0.601022.70/0.631424.85/0.7906
SAN [15]16M269G827.22/0.784025.14/0.647624.88/0.601022.70/0.631424.85/0.7906
PA-EDSR [14]45.53M872G 27.05/0.779625.05/0.645624.93/0.603422.85/0.639324.97/0.7915
DRN [40]5.4M171G 27.27/0.787325.21/0.650024.96/0.604022.93/0.638525.21/0.8016
NLSN [13]46.52M738G 26.57/0.758624.60/0.629924.66/0.592022.14/0.601124.00/0.7546
GPSR14.42M382G 27.45/0.793525.29/0.653625.00/0.605923.08/0.647425.40/0.8078
GPSR+14.42M- 27.52/0.795625.38/0.655525.05/0.607023.23/0.651825.64/0.8116
Table 5. Quantitative metrics obtained using various models for ×8 SR task with BD degradation model. The optimal and suboptimal metrics are highlighted in red and blue.
Table 5. Quantitative metrics obtained using various models for ×8 SR task with BD degradation model. The optimal and suboptimal metrics are highlighted in red and blue.
MethodParamsFLOPsScaleSet5Set14BSD100Urban100Manga109
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
DRN [40]5.4M171G 26.17/0.770424.43/0.639424.26/0.591622.22/0.631523.70/0.7782
RCAN [29]15.74M264G 26.22/0.771324.51/0.641524.28/0.593322.34/0.640223.87/0.7806
HAN [12]16.22M271G 26.20/0.761224.43/0.634624.28/0.588822.33/0.628623.71/0.7629
PA-EDSR [14]45.53M1271G826.26/0.772724.51/0.640124.29/0.593522.57/0.646023.85/0.7780
NLSN [11]46.52M738G 25.91/0.754424.04/0.624524.10/0.584021.80/0.606223.00/0.7465
GPSR14.42M382G 26.32/0.778824.60/0.646424.36/0.597422.85/0.660724.25/0.7950
GPSR+14.42M- 26.42/0.780924.72/0.649124.42/0.598723.05/0.665624.50/0.7993
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, X.; Huang, D.; Li, X.; Cai, D.; Zhu, D. GPSR: Gradient-Prior-Based Network for Image Super-Resolution. Appl. Sci. 2023, 13, 833. https://doi.org/10.3390/app13020833

AMA Style

Zhu X, Huang D, Li X, Cai D, Zhu D. GPSR: Gradient-Prior-Based Network for Image Super-Resolution. Applied Sciences. 2023; 13(2):833. https://doi.org/10.3390/app13020833

Chicago/Turabian Style

Zhu, Xiancheng, Detian Huang, Xiaorui Li, Danlin Cai, and Daxin Zhu. 2023. "GPSR: Gradient-Prior-Based Network for Image Super-Resolution" Applied Sciences 13, no. 2: 833. https://doi.org/10.3390/app13020833

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop