Next Article in Journal
Dual Solutions of Unsteady Mixed Convection Hybrid Nanofluid Flow Past a Vertical Riga Plate with Radiation Effect
Next Article in Special Issue
Survey of Cross-Modal Person Re-Identification from a Mathematical Perspective
Previous Article in Journal
A Complete Procedure for a Constraint-Type Fictitious Time Integration Method to Solve Nonlinear Multi-Dimensional Elliptic Partial Differential Equations
Previous Article in Special Issue
Resource- and Time-Efficient Computation Offloading in Vehicular Edge Computing: A Max-Min Fairness Oriented Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Single Image Super-Resolution Reconstruction with Preservation of Structure and Texture Details

1
Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
2
Faculty of Electrical Engineering, Kunming University of Science and Technology, Kunming 650500, China
3
Department of Computer Information Systems, State University of New York at Buffalo State, Buffalo, NY 14222, USA
4
School of Natural and Computational Sciences, Massey University at Auckland, Auckland 0632, New Zealand
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(1), 216; https://doi.org/10.3390/math11010216
Submission received: 29 November 2022 / Revised: 23 December 2022 / Accepted: 27 December 2022 / Published: 1 January 2023

Abstract

:
In recent years, deep-learning-based single image super-resolution reconstruction has achieved good performance. However, most existing methods pursue a high peak signal-to-noise ratio (PSNR), while ignoring the quality of the structure and texture details, resulting in unsatisfactory performance of the reconstruction results in terms of human subjective perception. To solve this issue, this paper proposes a structure- and texture-preserving image super-resolution reconstruction method. Specifically, two different network branches are used to extract features for image structure and texture details. A dual-coordinate direction perception attention (DCDPA) mechanism is designed to highlight structure and texture features. The attention mechanism fully considers the complementarity and directionality of multi-scale image features and effectively avoids information loss and possible distortion of image structure and texture details during image reconstruction. Additionally, a cross-fusion mechanism is designed to comprehensively utilize structure and texture information for super-resolution image reconstruction, which effectively integrates the structure and texture details extracted by the two branch networks. Extensive experiments verify the effectiveness of the proposed method and its superiority over existing methods.

1. Introduction

Image super-resolution (SR) focuses on recovering the corresponding high-resolution images from degraded low-resolution images. Image SR techniques can be used in face recognition [1], medical imaging [2,3], satellite imagery [4] and so on. As an important image-processing method, image SR has received extensive attention from researchers. Existing SR methods can be roughly divided into four categories: interpolation-based methods [5,6,7,8], inverse reconstruction-based methods [9,10,11,12], traditional machine-learning-based methods [13,14,15,16,17,18,19,20,21,22], and deep-learning-based methods [23,24,25,26,27,28,29,30,31].
Interpolation-based methods use known adjacent pixel information to generate unknown pixels. This type of method is simple to implement, but its detail recovery ability is poor, and the reconstruction results are prone to blur and aliasing. Assuming that the low-resolution image is obtained by a series of degradations of the high-resolution image, inverse reconstruction-based methods use the reverse reconstruction algorithm to restore the high-resolution image. Based on the degradation model of the image, this kind of method is often developed in combination with prior knowledge of the image [9,11] and obtains a high-resolution image by optimizing the model. Although this kind of method preserves more image details, the reconstruction performance is susceptible to image degradation and regularization performance. Additionally, when the resolution is improved by more than four times, the reconstruction performance is usually not good. Traditional learning-based methods use the training data to learn the non-linear mapping relationship between a low-resolution image and its high-resolution version. Then, the learned mapping relationship is used to guide high-resolution image reconstruction. Freeman et al. [32] first proposed the learning of this relationship in Markov neural networks, but this method requires a lot of time to construct a training set and image reconstruction is also time-consuming. Chang et al. [33] assumed that both high- and low-resolution image patches have manifolds with the same geometric structure and proposed a super-resolution reconstruction method based on neighborhood embedding. Since dictionary learning has shown excellent performance in computer vision [14,34] and medical imaging [35,36], Yang et al. [37] proposed a super-resolution reconstruction method based on sparse representation. This method assumes that the high- and low-resolution images have the same coding coefficients under the respective dictionary, and reconstructs the high-resolution image by applying the coding coefficients of the low-resolution image under the low-resolution dictionary to the high-resolution dictionary. In view of the effectiveness of sparse representation and its superiority for image processing [38,39,40,41] and recognition [42,43,44], image super-resolution reconstruction based on sparse representation became a very popular method at that time. Although this type of method has achieved breakthroughs in performance, the optimization process of the model is too complicated, time-consuming and memory-intensive.
In 2014, Dong et al. [17] first applied deep learning to image super-resolution reconstruction and proposed super-resolution convolutional neural networks (SRCNN). Since then, deep-learning-based methods have become the mainstream methods for image super-resolution reconstruction [23,24,25,26,27,28,29,30,31,45,46]. This kind of method usually uses a single-branch network under the constraint of loss function so that the network can extract the texture details required to restore the low-resolution image to the high-resolution image, thereby effectively improving the visual effect of the reconstructed image. However, texture features are easily lost in the process of image convolution in real cases. This is extremely disadvantageous for recovering texture details in high-resolution images. In order to improve the visual performance of reconstructed images, Ledig et al. [13] proposed super-resolution generative adversarial networks (SRGAN), which use the generative adversarial loss to constrain the generator and discriminator. Hence, the generator obtains texture details that are closer to the real world. Since then, Wang et al. [47] further improved the detail recovery ability of SRGAN by introducing dense residual blocks. However, the discriminator may introduce instability in the optimization process during the adversarial process, causing structure inconsistency in the reconstructed image, resulting in unnatural artifacts and geometric distortions in the recovered super-resolution image. Therefore, Ma et al. [24] proposed a structure-preserving super-resolution (SPSR) method to alleviate the above problem, which uses the gradient information of the image to guide the low-resolution image restoration, while maintaining the advantages of GAN-based methods to generate results that achieve satisfactory performance in terms of human subjective perception.
In order to reconstruct the structure and texture details of images with high quality, this paper proposes a structure- and texture-preserving image super-resolution method. In the proposed method, the structure and texture details of the image are extracted by different network branches. In this process, a dual-coordinate direction perception attention (DCDPA) mechanism is designed to highlight structure and texture features. This mechanism not only makes full use of the complementarity of multi-scale image features, but also considers the directionality of image features in the horizontal and vertical coordinates. Therefore, both information loss and possible distortion of the image structure and texture details are effectively avoided during image reconstruction. To comprehensively utilize the structure and texture information to reconstruct high-resolution images, this paper proposes a cross-fusion mechanism to integrate structure and texture features.
In the network structure, the proposed network framework is mainly composed of a structure feature extraction module (SFEM), a texture detail feature extraction module (TDFEM), and feature highlight and aggregation modules (FHAM). The SFEM and TDFEM are responsible for the learning of image structure features and texture detail features, respectively. The FHAM enhances the structure and texture detail features in different directions in the image through the designed DCDPA mechanism and promotes information interaction between them through the cross fusion of the structure and texture detail features to improve the feature representation ability. The network is optimized by reconstruction loss, texture loss, and gradient loss. This paper makes three main contributions as follows:
  • An image super-resolution reconstruction network for structure and texture restoration is designed. This network uses the dual-stream structure to extract the structure and texture details of the image, respectively, and enhances these two features layer-by-layer through the FHAM to improve the network’s ability to extract features.
  • The FHAM is designed to highlight the structure and texture features in the image. In order to enable the network to capture the salient structure and texture features in the image, this paper proposes a DCDPA mechanism in this module. This mechanism makes full use of the directionality of image features to highlight the features of different directions in the image. Moreover, a cross-fusion block (CFB) is proposed to promote the interactive fusion of structure and texture features.
  • A large number of experimental results show that the proposed method contains salient structure and texture details in its super-resolution reconstruction results, which outperform state-of-the-art methods in visual quality and performance evaluation.

2. Relate Work

2.1. Deep-Learning-Based Super-Resolution Reconstruction Methods

In recent years, deep learning has been widely used in single-image super-resolution reconstruction. Dong et al. [17] proposed the SRCNN. Compared with traditional methods, SRCNN achieves better super-resolution reconstruction results. Since simply stacking convolutional layers can lead to gradient explosion, Kim et al. [22] proposed an accurate image super-resolution using very deep convolutional networks (VDSR). The networks avoid the problem of gradient explosion and speed up the network training by adding jump connections. DRRN [48] and Memnet [49] proposed by Tai et al. introduced recursive blocks and memory blocks, respectively, to stack more convolutional layers. However, the above-mentioned networks all use predefined upsampling. Due to resolution changes using traditional methods, the network complexity increases. In addition, since the traditional upsampling methods do not involve network parameter training, the image cannot be reconstructed by convolution, activation function, etc. Therefore, the upsampling process magnifies the defects in low-resolution images. To solve this problem, Dong et al. [25] proposed a fast convolutional neural-network-based super-resolution reconstruction (FSRCNN) method, which performs feature extraction on low-resolution images, reduces the complexity of the network, and applies deconvolution to upsampling and image reconstruction. Shi et al. [26] proposed an efficient sub-pixel convolutional layer to achieve upsampling. In recent years, Zhu et al. [27] proposed a lightweight feature separation and fusion network, adopting a block-processing structural framework, which enables the network to fully utilize and enrich features at different levels. Liu et al. [28] proposed a multi-scale skip-connection network (MSN) to improve the visual quality of super-resolution images, utilizing convolution kernels of different sizes to capture multi-scale features of low-resolution images. Huang et al. [29] proposed an attention network with detail fidelity, which achieves super-resolution image reconstruction while obtaining detail fidelity. Mehri et al. [30] proposed the multi-path residual network (MPRNet), which adaptively learns the most valuable features to learn more high-frequency information about the image. Tan et al. [31] proposed an image super-resolution method via a self-calibrated feature fusion network (SCFFN) to achieve a better balance between network performance and applicability. However, the above methods all consider the texture feature information and structure feature information of the image as a whole. In reality, the texture features are easily lost during the image convolution process. This is extremely disadvantageous for reconstructing high-resolution images rich in texture feature information.

2.2. Visual-Perception-Based Super-Resolution Reconstruction Methods

To improve the perceptual quality of images, in recent years, a large number of super-resolution reconstruction models that benefit the perceptual performance of human eyes have been proposed [13,47,50]. These methods use a variety of information compensation methods to compensate the texture features of the super-resolution reconstruction results. While reconstructing the structure information, they prevent the loss of texture details during the super-resolution reconstruction process and improve the perceptual performance of images. A photo-realistic single image super-resolution using a generative adversarial network (SRGAN) proposed by Ledig et al. [13] uses the discriminator and the generator for adversarial learning and enables the generator to generate texture details that are close to the real world. In order to reduce the artifacts caused by generative confrontation in SRGAN, Wang et al. [47] proposed an enhanced super-resolution generative adversarial network (ESRGAN), which introduced dense residual blocks into the basic network framework and abandoned the batch normalization operation to generate more realistic and natural images. Wu et al. [51] proposed a perceptual generative adversarial network for single-image super-resolution (SRPGAN), in which a robust perceptual loss based on the model’s discriminator was proposed to recover texture details. Wang et al. [50] proposed an image super-resolution reconstruction method for texture recovery by deep spatial feature transform, which integrates a semantic prior into low-resolution images and uses a spatial feature transform (SFT) layer to further improve texture recovery. A new dense block proposed by Chen et al. [52] uses complex connections between each layer to build a more powerful generator and applies new feature maps to compute the perceptual loss to make the output image more realistic and natural. Ma et al. [24] proposed a structure-preserving image super-resolution method, which uses the gradient information of the image to guide image recovery. Fu et al. [53] proposed a method to guide the image super-resolution reconstruction by weak texture information. This network consists of a main network and two auxiliary prediction networks. The main network extracts and combines the main distinct depth features to assist the prediction network in extracting weak texture information. Cai et al. [54] proposed a texture and detail preservation network (TDPN), which not only focuses on local region feature recovery, but also texture and detail preservation. Meng et al. [55] proposed a gradient information distillation network that not only maintains the advantage of fast and lightweight, but also improves the model performance through information distillation. All of the above-mentioned methods have made great contributions to improving the visual perception of the reconstructed images. However, unnatural artefacts still appear in the recovered super-resolution images. Therefore, this paper proposes a super-resolution reconstruction network that fuses structure and texture information. By enhancing the structure and texture features in the image and cross-fusion, the extraction ability of the network for structure and texture features is improved, thereby reducing artefacts in the reconstructed image and improving the perceptual quality of the reconstructed image.

3. The Proposed Network Framework

The framework of the proposed image super-resolution reconstruction network for structure and texture detail recovery is shown in Figure 1. The model mainly comprises three modules: SFEM, TDFEM, and FHAM. The network structure of both SFEM and TDFEM is the same, but the parameters are not shared. They both consist of a 5 × 5 convolutional layer and five sets of residual blocks (RBs). Convolutional layers and RBs extract shallow and deep features of the input image, respectively. The input of the SFEM is the original low-resolution image. Additionally, the texture map of the original low-resolution image is obtained by the local binary pattern (LBP) method [56]; the texture map and the original image are concatenated as the input of the TDFEM. The FHAM consists of five gated fusion blocks (GFBs). The GFB enhances the structure and texture information of the image, respectively, by introducing a DCDPA mechanism and fusing the enhanced structure and texture information through cross-fusion. Finally, the reconstructed super-resolution image is obtained by deconvolution.

3.1. Structural Feature Extraction Module and Texture Detail Feature Extraction Module

The network structure of both SFEM and TDFEM is the same, including shallow-feature extraction and deep-feature extraction. The network adopts a 5 × 5 convolution layer and five RBs to extract the shallow and deep features of the image, respectively. Assuming that I represents the input of the 5 × 5 convolutional layer, the shallow feature F s obtained by convolution is:
F s = s c o n v ( I , k = 5 )
where s c o n v represents equal-sized convolution operations, and k represents the size of the convolution kernel.
Let I L R denote the input original low-resolution image. For the SFEM I = I L R , the obtained shallow structure feature is F s s . In order to prevent the subjective image quality degradation caused by loss of texture detail information during the super-resolution reconstruction process, the LBP method is used to obtain the texture map I t of the original image I L R and to concatenate them with the input of the TDFEM. Therefore, for the TDFEM, I = c o n c a t (   I L R , I t ) , where c o n c a t denotes the concatenation operation. The obtained shallow texture detail feature is F t s . The RB is used to extract the deep features of the image. The relationship between the input and output of the i-th ( i = 1 , 2 , , 5 ) RB can be expressed as:
F r i = s c o n v g ( F r i 1 ) + F r i 1
where s c o n v g represents a convolution group, which consists of six 3 × 3 convolutional layers. F r i 1 and F r i represent the input and output of the i-th RB, respectively.

3.2. Feature Highlight and Aggregation Module and Feature Reconstruction

To ensure that the network can realize the fusion of image structure and texture features and prevent the loss of structure and texture details, this paper proposes the FHAM. The module consists of five GFBs, as shown in Figure 1. The GFB is composed of DCDPA and CFB.

3.2.1. Dual-Coordinate Direction Perception Attention

To enable the network to capture structure and texture details in different directions from the image, a DCDPA mechanism is used, which includes direction perception coordinate attention (DPCA) and global perception coordinate attention (GPCA), as shown in Figure 2.
The input F G i n of the DCDPA module contains the texture detail and structure features of the image. Specifically, for the first GFB, its input is obtained by concatenating the shallow structure feature F s s and the shallow texture detail feature F t s . For other GFBs, the corresponding input is the output of the previous GFB. F G i n passes through 3 × 3 and 5 × 5 convolutions, respectively, to obtain the features F l ( l = 3 , 5 ) of different scales. F l obtains the horizontal feature F l h and the vertical feature F l v through global average pooling (GAP) in the horizontal and vertical directions, respectively:
F l h = P h ( F l ) F l v = P v ( F l )
where l = 3 , 5 , and P h and P v represent the GAP in the horizontal and vertical directions, respectively. Set F l R 3 × H × W , then F l v R 3 × 1 × W , F l h R 3 × H × 1 . To concatenate features in different directions, it is necessary to reshape F l h into 3 × 1 × H dimensions. The features in different directions are concatenated along the horizontal direction to obtain four sets of joint features as follows:
F 3 , 3 h v = c o n c a t ( F 3 h , F 3 v ) F 3 , 5 h v = c o n c a t ( F 3 h , F 5 v ) F 5 , 3 h v = c o n c a t ( F 5 h , F 3 v ) F 5 , 5 h v = c o n c a t ( F 5 h , F 5 v )
where c o n c a t represents the concatenation operation. To obtain the perceptual coordinate attention weights in different directions, 1 × 1 convolution and s i g m o i d are applied to process the joint features to obtain attention maps in different directions. Taking F 3 , 3 h v as an example, the attention maps in the horizontal and vertical directions obtained after F 3 , 3 h v is processed are M 3 , 3 h R 3 × 1 × W and M 3 , 3 v R 3 × 1 × W :
M 3 , 3 h , M 3 , 3 v = s p l i t ( s i g m o i d ( B N ( s c o n v ( F 3 , 3 h v , k = 1 ) ) ) )
where s p l i t represents the segmentation operation, s i g m o i d represents the sigmoid activation function, and B N represents the batch normalization layer. After applying 1 × 1 convolution and dimension expansion in the horizontal and vertical directions to M 3 , 3 h and M 3 , 3 v , the perceptual coordinate attention maps M ˜ 3 , 3 h R 3 × W × H and M ˜ 3 , 3 v R 3 × W × H in different directions are obtained, respectively. Similarly, other attention maps M ˜ 3 , 5 h , M ˜ 3 , 5 v , M ˜ 5 , 3 h , M ˜ 5 , 3 v , and M ˜ 5 , 5 h , M ˜ 5 , 5 v , can be obtained.
The features in different directions in an image are enhanced using DPCA in both the vertical and horizontal directions. F 3 is enhanced with attention maps M ˜ 3 , 3 h , M ˜ 3 , 3 v and M ˜ 3 , 5 h , M ˜ 3 , 5 v , respectively, and the enhanced features are F ˜ 3 , 3 and F ˜ 3 , 5 :
F ˜ 3 , 3 = F 3 M ˜ 3 , 3 h M ˜ 3 , 3 v F ˜ 3 , 5 = F 3 M ˜ 3 , 5 h M ˜ 3 , 5 v
where ⊙ denotes point-wise multiplication.
Similarly, F 5 is enhanced with attention maps M ˜ 5 , 3 h , M ˜ 5 , 3 v and M ˜ 5 , 5 h , M ˜ 5 , 5 v , respectively, and the enhanced features are F ˜ 5 , 3 and F ˜ 5 , 5 , respectively. The GPCA map M a t t is generated through the concatenation, convolution and sigmoid activation function of the features F ˜ 3 , 3 , F ˜ 3 , 5 , F ˜ 5 , 3 and F ˜ 5 , 5 :
M a t t = s i g m o i d ( s c o n v ( c o n c a t ( F ˜ 3 , 3 , F ˜ 3 , 5 , F ˜ 5 , 3 , F ˜ 5 , 5 ) ) )
The output F s i of the i-th RB of the structure feature extraction module is enhanced by the GPCA map M a t t i ( i = 1 , 2 , , 5 ) in the i-th GFB, and the output F t i of the i-th RB of the TDFEM is enhanced by the complementary attention map 1 M a t t i . The enhanced structure feature and texture feature are F s e i and F t e i , respectively, as follows:
F s e i = F s i M a t t i F t e i = F t i ( 1 M a t t i )

3.2.2. Cross-Fusion Block

To make full use of the structure and texture detail information enhanced by the DCDPA mechanism, a cross-fusion block (CFB) is proposed, as shown in Figure 3. Specifically, the structural feature F s e and texture feature F t e generate the spatial attention map M s through concatenation, convolution and the sigmoid:
M s = s i g m o i d ( s c o n v ( c o n c a t ( F s e , F t e ) , k = 3 ) )
Using M s to further enhance the structure features, F s t is obtained by the cross-fusion of the enhanced structure features and texture features. Similarly, F t s is obtained by the cross-fusion of the structure features and the enhanced texture results by M s :
F s t = F s e M s + F t e F t s = F t e M s + F s e
The final fusion result is obtained by the concatenation of F s t and F t s :
F G o u t = c o n c a t ( F s t , F t s )
The fusion result F G o u t i of the i-th GFB is used as the input of the i + 1 -th GFB, i.e., F G i n i + 1 = F G o u t i .

3.2.3. Feature Reconstruction

To obtain the reconstructed image, the output F G o u t d of the last layer in the FHAM and the output F t d of the last layer in the TDFEM are upsampled by deconvolution to obtain F h and F t h , respectively. The texture detail feature map F t h is concatenated with F h to obtain features with further enhanced details; the high-resolution image I S R is finally reconstructed by 3 × 3 convolution:
F h = D e c o n v ( F G o u t d ) F t h = D e c o n v ( F t d ) I S R = s c o n v ( c o n c a t ( F h , F t h ) , k = 3 )
where D e c o n v stands for deconvolution.

3.3. Loss Function

The model is trained by a joint loss including reconstruction loss, gradient loss, and texture loss to achieve good visual performance.
Reconstruction loss: In this paper, the labeled high-resolution image, and the high-resolution image after super-resolution reconstruction, are constrained by the reconstruction loss as follows:
L r e = I H R I S R 1
where I H R and I S R are the labeled high-resolution image and reconstructed super-resolution image, respectively.
Gradient loss: To ensure that the reconstruction result has the same gradient information as the labeled image, the following gradient loss function to further optimize the reconstructed image is used.
L g r a d = 1 H × W I H R I S R 1
where ∇ is the gradient operator and H and W represent the height and width of the image, respectively.
Texture loss: To make the texture features in the reconstructed image and the labeled image consistent, a texture loss is introduced as follows:
L t e x = I H R t I S R t 1
where I H R t and I S R t are the texture image extracted from the labeled high-resolution image by the LBP operator and the texture image reconstructed by the texture detail feature extraction module, respectively. So, the joint loss is summarized as follows:
L j o i n t = L r e + λ g r a d L g r a d + λ t e x L t e x
where λ g r a d and λ t e x represent the weight parameters.

4. Experiment and Result Analysis

4.1. Experimental Setup

The Pytorch 1.7 framework and NVIDIA GeForce RTX 2080ti 12 GB GPU were used in the experiments. The training images were from the DIV2K dataset [57], which includes 800 training images, 100 validation images, and 100 test images. Four benchmark datasets, Set5 [58], Set14 [59], B100 [60], Urban100 [61], were used to evaluate the performance of the different methods. The Set5, Set14, and B100 datasets contained 5, 10, and 100 natural images, respectively. There were 100 urban building images in the Urban100 dataset. The high-resolution testing images in the testing set were downsampled by 2 × , 3 × and 4 × bicubic interpolation to obtain the corresponding low-resolution images for super-resolution reconstruction testing. When training the network model with a 2 × upsampling factor, to reduce memory usage and save running time, the size of the high-resolution image was 256 × 256 and the size of the low-resolution image was 128 × 128 . When training the network model with a 3 × upsampling factor, the size of the high-resolution image was 192 × 192 , and the size of the low-resolution image was 64 × 64 . When training the network model with a 4 × upsampling factor, the size of the high-resolution image was 180 × 180 and the size of the low-resolution image was 45 × 45 . In the training process, the Adam optimizer [62] was selected to train all modules. Following [54], the learning rate of the model training was set to 10 4 . The convergence curves of the training loss during the training for scale 2 magnification on the DIV2K dataset are shown in Figure 4. It can be seen from Figure 4 that the training loss gradually decreased as the training epochs increased. When it reached 1800 epochs, the loss curve converged. Allowing for some margin, the epoch for the model training was set to 2000. The parameters of the entire network were trained through back propagation.
Two objective evaluation indicators, the peak signal-to-noise ratio (PSNR) and a structural similarity index measure (SSIM) were used to evaluate the quality of the super-resolution reconstruction results. The PSNR was used to measure the mean-square error between the reconstructed image and the original image; its unit is dB. The larger the value of PSNR, the less the image distortion, that is, the higher the image reconstruction quality. The SSIM was used to measure the structural similarity between the reconstructed image and the original image. A higher value of SSIM indicates that the reconstructed image is closer to the original image, that is, it shows better image reconstruction performance.
Given two images, one being the ground truth (GT) X , the other the super-resolution reconstruction result Y , the PSNR [63] can be defined as:
P S N R = 10 log 10 ( ( 2 n 1 ) 2 M S E )
where M S E is the mean square error of the GT image X and the reconstructed image Y . M S E = 1 H × W i = 1 H j = 1 W ( X ( i , j ) Y ( i , j ) ) 2 . H and W are the height and width of the image, respectively. n is the bits of a pixel which is generally set to 8 (i.e., a gray level of 256). The unit of the PSNR is the decibel (dB).
The SSIM [64] is defined as:
S S I M ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 )
where μ x and μ y are the mean values of X and Y , respectively. σ x and σ y are the standard deviations of X and Y , respectively. σ x y is the covariance of X and Y . c 1 and c 2 are set to constants to avoid system error caused by a zero denominator. The value of the SSIM is between 0 and 1.

4.2. Quantitative Comparison

To verify the effectiveness of the proposed method, it was compared with state-of-the-art super-resolution methods, such as SRCNN [65], FSRCNN [25], VDSR [22], LapSRN [66], CDC [67], and SREFBN [68]. SRCNN and FSRCNN were trained on the 91-image dataset [17]. VDSR, LapSRN, CDC and SREFBN were trained on the DIV2K dataset. The above methods were all validated on the Set5, Set14, B100, and Urban100 datasets. The quantitative evaluation results are shown in Table 1, Table 2 and Table 3 with the best results marked in bold. According to Table 1, Table 2 and Table 3, the proposed method achieved most of the best evaluation results.

4.3. Qualitative Comparison

To compare the visual quality of the super-resolution reconstruction results obtained by different methods, images from the Set5, Set14, BSD100, and Urban100 datasets were selected for the visual display of the 2 × , 3 × , and 4 × super-resolution reconstruction results and ground truth (GT), as shown in Figure 5, Figure 6 and Figure 7. In terms of visual performance, the detailed textures of the reconstructed images obtained by the SRCNN, FSRCNN, VDSR, and CDC methods were relatively blurred, while the reconstruction results obtained by the LapSRN, CDC, and SREFBN methods contained detailed features that did not match the actual texture details. The proposed method could fit real natural images well, enhancing the brightness of reconstructed high-resolution images, while recovering textures, so that the reconstructed images had rich texture details. In summary, the proposed method has certain advantages in terms of visual performance and quantitative evaluation.

4.4. Ablation Study

To verify the effectiveness of the proposed TDFEM, DCDPA, and CFB methods on the quality of the reconstructed super-resolution images, ablation experiments were performed. Taking the network without DCDPA, CFB and TDFEM as the benchmark (Base), the Base network plus TDFEM was marked as Base + TDFEM, the Base + TDFEM network plus DCDPA was marked as Base + TDFEM + DCDPA, and the Base + TDFEM + DCDPA network plus CFB was marked as Base+TDFEM+DCDPA+CFB. Ablation experiments were performed on the images from the Urban100 dataset at 2 × magnification. The quantitative evaluation results of the ablation experiments are shown in Table 4 and the corresponding visual performance is shown in Figure 8.

4.4.1. The Effectiveness of TDFEM

TDFEM was used to extract texture information separately to prevent the loss of texture detail information in the process of super-resolution reconstruction, which led to the deterioration in subjective image quality. So, TDFEM was introduced into the proposed model. In the TDFEM, the LBP operator is used to extract the texture details of the original image and the obtained texture detail map is concatenated with the original image, which enhances the network’s ability to extract texture detail information. This helps to improve the perceptual quality of reconstructed super-resolution images. To demonstrate its effectiveness, the reconstruction results of the ‘Base’ model were compared with the reconstruction results of the ‘Base + TDFEM’ model. As shown in Figure 8, the ‘Base + TDFEM’ model achieved good performance in texture detail extraction during the reconstruction process. Additionally, as shown in Table 4, the utilization of ‘TDFEM’ also led to a significant improvement in the objective evaluation results, which confirmed the effectiveness of ‘TDFEM’ in the proposed model.

4.4.2. The Effectiveness of DCDPA

In the proposed model, DCDPA is used to capture the structure and texture feature regions of interest and to prevent the loss of feature information. This attention mechanism makes full use of the complementarity and directionality of multi-scale image features and effectively avoids information loss and the possible distortion of image structure and texture details during image reconstruction. This is conducive to improving the visual quality of the reconstructed image. To verify the effectiveness of ‘DCDPA’, the reconstruction results of the ‘Base + TDFEM + DCDPA’ model were compared with those of the ‘Base + TDFEM’ model. In Figure 8, the structure and texture details of the reconstructed images obtained by the ‘Base + TDFEM + DCDPA’ are highlighted. Moreover, according to the values of the objective evaluation metrics shown in Table 4, compared with ‘Base + TDFEM’, the performance of ‘Base + TDFEM + DCDPA’ was improved. So, the effectiveness of DCDPA was demonstrated.

4.4.3. The Effectiveness of CFB

To prevent the loss of texture details and ensure that the texture detail information enhanced by the DCDPA was fully utilized by the later layers, a CFB was proposed. This module facilitates the information interaction between structure and texture detail features by cross-fusion of them to enhance the network’s ability to mine structure and texture features. This contributes to improving the quality of the reconstructed image. To verify the effectiveness of ‘CFB’, the reconstruction results of ‘Base + TDFEM + DCDPA + CFB’ were compared with those of ‘Base + TDFEM + DCDPA’. As shown in Figure 8, the ‘Base + TDFEM + DCDPA + CFB’ model more fully integrated structure and texture features in the reconstruction process and the corresponding evaluation results presented in Table 4 were also enhanced.

4.5. Parameter Analysis

4.5.1. Influence of the Number of GFBs

This section discusses the influence of the number of GFBs on the performance of the proposed model. The proposed model with different numbers of GFBs was tested on the DIV2K dataset. Table 5 shows the performance of the model with different numbers of GFBs (denoted by n). ’Params’ indicates the number of parameters. As shown in Table 5, PSNR and SSIM achieved the best performance at n = 5 . When n was 7 or 9, the number of model parameters increased and the model’s performance decreased. Therefore, n was set to 5.

4.5.2. Hyper-Parameter Selection

In Equation (16), two hyper-parameters λ g r a d and λ t e x need to be set. The impact of one hyper-parameter is analyzed by fixing another hyper-parameter. For the DIV2K dataset, the impact of λ g r a d and λ t e x on the super-resolution reconstruction results was analyzed by PSNR and SSIM, respectively; the corresponding experimental results are shown in Figure 9. As shown in Figure 9, when λ g r a d = 0.01 and λ t e x = 0.001 , both PSNR and SSIM achieved the optimal values. Therefore, λ g r a d = 0.01 and λ t e x = 0.001 were set for model training.

5. Conclusions

A super-resolution reconstruction network that preserves structure and texture details was designed. The proposed network uses different network branches for feature extraction of the image structure and texture details, respectively, and enhances these two features layer-by-layer through the FHAM to improve the network’s ability to extract features. To fully extract the information of different directions in the image, a dual-coordinate direction perception attention mechanism is proposed. The proposed attention mechanism fully considers the complementarity and directionality of the multi-scale image features, which can avoid information loss and possible distortion of the image structure and texture details. Additionally, a cross-fusion mechanism to effectively integrate the structure and texture features extracted by the network was designed, which can enhance the network’s ability to mine structure and texture features. This is helpful to improve the quality of the reconstructed image. The experimental results show that the super-resolution reconstruction results of the proposed method produced good structure and texture details. The proposed method outperformed state-of-the-art methods in both visual quality and performance evaluation. However, the proposed method uses synthetic datasets to train networks in a supervised way. In the synthetic datasets, a low-resolution image is generated by bicubic downsampling its high-resolution counterpart. In practical scenarios, the high-resolution images are unavailable and image degradation models are unknown. Therefore, how to improve the quality of super-resolution reconstructed images in real scenes will be the direction of our future research.

Author Contributions

Conceptualization, K.W. and Y.H.; methodology, Y.Z. and Y.H.; software, Y.H.; validation, J.Z.; formal analysis, Y.H.; investigation, Y.Z. and Y.H.; data curation, Y.H.; writing—original draft preparation, Y.Z., Y.H. and G.Q.; writing—review and editing, Y.Z., Y.H. and G.Q.; visualization, Y.H. and J.Z.; supervision, K.W.; project administration, Y.Z. and K.W.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China No. 62161015.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Program and test data are available at https://github.com/YafeiZhang-KUST/PSTD, accessed on 28 November 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. ElSayed, A.; Mahmood, A.; Sobh, T. Effect of Super Resolution on High Dimensional Features for Unsupervised Face Recognition in the Wild. In Proceedings of the 2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 10–12 October 2017. [Google Scholar]
  2. Zhao, X.; Zhang, Y.; Zhang, T.; Zou, X. Channel splitting network for single MR image super-resolution. IEEE Trans. Image Process. 2019, 28, 5649–5662. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Li, H.; Yu, Z.; Mao, C. Fractional differential and variational method for image fusion and super-resolution. Neurocomputing 2016, 171, 138–148. [Google Scholar] [CrossRef]
  4. Meishvili, G.; Jenni, S.; Favaro, P. Learning to have an ear for face super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1364–1374. [Google Scholar]
  5. Keys, R. Cubic convolution interpolation for digital image processing. IEEE Trans. Acoust. Speech Signal Process. 1981, 29, 1153–1160. [Google Scholar] [CrossRef] [Green Version]
  6. Zhang, L.; Wu, X. An edge-guided image interpolation algorithm via directional filtering and data fusion. IEEE Trans. Image Process. 2006, 15, 2226–2238. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Sanchez-Beato, A.; Pajares, G. Noniterative interpolation-based super-resolution minimizing aliasing in the reconstructed image. IEEE Trans. Image Process. 2008, 17, 1817–1826. [Google Scholar] [CrossRef]
  8. Zhou, F.; Yang, W.; Liao, Q. Interpolation-based image super-resolution using multisurface fitting. IEEE Trans. Image Process. 2012, 21, 3312–3318. [Google Scholar] [CrossRef]
  9. Sun, J.; Xu, Z.; Shum, H.Y. Image super-resolution using gradient profile prior. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AL, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  10. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2272–2279. [Google Scholar]
  11. Tai, Y.W.; Liu, S.; Brown, M.S.; Lin, S. Super resolution using edge prior and single image detail synthesis. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2400–2407. [Google Scholar]
  12. Zhang, K.; Gao, X.; Tao, D.; Li, X. Single image super-resolution with non-local means and steering kernel regression. IEEE Trans. Image Process. 2012, 21, 4544–4556. [Google Scholar] [CrossRef]
  13. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 105–114. [Google Scholar]
  14. Timofte, R.; De Smet, V.; Van Gool, L. Anchored neighborhood regression for fast example-based super-resolution. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1920–1927. [Google Scholar]
  15. Chen, C.L.P.; Liu, L.; Chen, L.; Tang, Y.Y.; Zhou, Y. Weighted couple sparse representation with classified regularization for impulse noise removal. IEEE Trans. Image Process. 2015, 24, 4014–4026. [Google Scholar] [CrossRef]
  16. Liu, L.; Chen, L.; Chen, C.L.P.; Tang, Y.Y.; Pun, C.M. Weighted joint sparse representation for removing mixed noise in image. IEEE Trans. Cybern. 2017, 47, 600–611. [Google Scholar] [CrossRef]
  17. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar]
  18. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  19. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  20. Liu, J.; Zhang, W.; Tang, Y.; Tang, J.; Wu, G. Residual feature aggregation network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2356–2365. [Google Scholar]
  21. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 191–207. [Google Scholar]
  22. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  23. Soh, J.W.; Park, G.Y.; Jo, J.; Cho, N.I. Natural and realistic single image super-resolution with explicit natural manifold discrimination. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8114–8123. [Google Scholar]
  24. Ma, C.; Rao, Y.; Cheng, Y.; Chen, C.; Lu, J.; Zhou, J. Structure-preserving super resolution with gradient guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 7766–7775. [Google Scholar]
  25. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 391–407. [Google Scholar]
  26. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  27. Zhu, K.; Chen, Z.; Wu, Q.; Wang, N.; Zhao, J.; Zhang, G. FSFN: Feature separation and fusion network for single image super-resolution. Multimed. Tools Appl. 2021, 80, 31599–31618. [Google Scholar] [CrossRef]
  28. Liu, J.; Ge, J.; Xue, Y.; He, W.; Sun, Q.; Li, S. Multi-scale skip-connection network for image super-resolution. Multimed. Syst. 2021, 27, 821–836. [Google Scholar] [CrossRef]
  29. Huang, Y.; Li, J.; Gao, X.; Hu, Y.; Lu, W. Interpretable detail-fidelity attention network for single image super-resolution. IEEE Trans. Image Process. 2021, 30, 2325–2339. [Google Scholar] [CrossRef] [PubMed]
  30. Mehri, A.; Ardakani, P.B.; Sappa, A.D. MPRNet: Multi-path residual network for lightweight image super resolution. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 2703–2712. [Google Scholar]
  31. Tan, C.; Cheng, S.; Wang, L. Efficient image super-resolution via self-calibrated feature fuse. Sensors 2022, 22, 329. [Google Scholar] [CrossRef] [PubMed]
  32. Freeman, W.T.; Pasztor, E.C.; Carmichael, O.T. Learning low-level vision. Int. J. Comput. Vis. 2000, 40, 25–47. [Google Scholar] [CrossRef]
  33. Chang, H.; Yeung, D.Y.; Xiong, Y. Super-resolution through neighbor embedding. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 27 June–2 July 2004; Volume 1, p. 1. [Google Scholar]
  34. Timofte, R.; De Smet, V.; Van Gool, L. A+: Adjusted anchored neighborhood regression for fast super-resolution. In Proceedings of the Asian Conference on Computer Vision, Singapore, 1–5 November 2014; pp. 111–126. [Google Scholar]
  35. Rahim, T.; Novamizanti, L.; Ramatryana, I.N.A.; Shin, S.Y.; Kim, D.S. Total variant based average sparsity model with reweighted analysis for compressive sensing of computed tomography. IEEE Access 2021, 9, 119158–119170. [Google Scholar] [CrossRef]
  36. Rahim, T.; Novamizanti, L.; Ramatryana, I.N.A.; Shin, S.Y. Compressed medical imaging based on average sparsity model and reweighted analysis of multiple basis pursuit. Comput. Med Imaging Graph. 2021, 90, 101927. [Google Scholar] [CrossRef]
  37. Yang, J.; Wright, J.; Huang, T.; Ma, Y. Image super-resolution as sparse representation of raw image patches. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AL, USA, 24–26 June 2008; pp. 1–8. [Google Scholar]
  38. Li, H.; He, X.; Tao, D.; Tang, Y.; Wang, R. Joint medical image fusion, denoising and enhancement via discriminative low-rank sparse dictionaries learning. Pattern Recognit. 2018, 79, 130–146. [Google Scholar] [CrossRef]
  39. Xie, M.; Wang, J.; Zhang, Y. A unified framework for damaged image fusion and completion based on low-rank and sparse decomposition. Signal Process. Image Commun. 2021, 29, 116400. [Google Scholar] [CrossRef]
  40. Zhang, Y.; Yang, M.; Li, N.; Yu, Z. Analysis-synthesis dictionary pair learning and patch saliency measure for image fusion. Signal Process. 2020, 167, 107327. [Google Scholar] [CrossRef]
  41. Li, H.; Wang, Y.; Yang, Z.; Wang, R.; Li, X.; Tao, D. Discriminative Dictionary Learning-Based Multiple Component Decomposition for Detail-Preserving Noisy Image Fusion. IEEE Trans. Instrum. Meas. 2020, 69, 1082–1102. [Google Scholar] [CrossRef]
  42. Li, H.; Yan, S.; Yu, Z.; Tao, D. Attribute-Identity Embedding and Self-Supervised Learning for Scalable Person Re-Identification. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 3472–3485. [Google Scholar] [CrossRef]
  43. Yan, S.; Zhang, Y.; Xie, M.; Zhang, D.; Yu, Z. Cross-domain person re-identification with pose-invariant feature decomposition and hypergraph structure alignment. Neurocomputing 2022, 467, 229–241. [Google Scholar] [CrossRef]
  44. Li, H.; Xu, J.; Yu, Z.; Luo, J. Jointly Learning Commonality and Specificity Dictionaries for Person Re-Identification. IEEE Trans. Image Process. 2020, 29, 7345–7358. [Google Scholar] [CrossRef]
  45. Li, H.; Cen, Y.; Liu, Y.; Chen, X.; Yu, Z. Different Input Resolutions and Arbitrary Output Resolution: A Meta Learning-Based Deep Framework for Infrared and Visible Image Fusion. IEEE Trans. Image Process. 2021, 30, 4070–4083. [Google Scholar] [CrossRef]
  46. Xiao, W.; Zhang, Y.; Wang, H.; Li, F.; Jin, H. Heterogeneous Knowledge Distillation for Simultaneous Infrared-Visible Image Fusion and Super-Resolution. IEEE Trans. Instrum. Meas. 2022, 71, 5004015. [Google Scholar] [CrossRef]
  47. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018; pp. 63–79. [Google Scholar]
  48. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2790–2798. [Google Scholar]
  49. Tai, Y.; Yang, J.; Liu, X.; Xu, C. Memnet: A persistent memory network for image restoration. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4549–4557. [Google Scholar]
  50. Wang, X.; Yu, K.; Dong, C.; Loy, C.C. Recovering realistic texture in image super-resolution by deep spatial feature transform. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 606–615. [Google Scholar]
  51. Wu, B.; Duan, H.; Liu, Z.; Sun, G. SRPGAN: Perceptual generative adversarial network for single image super resolution. arXiv 2017, arXiv:1712.05927. [Google Scholar]
  52. Chen, B.X.; Liu, T.J.; Liu, K.H.; Liu, H.H.; Pei, S.C. Image super-resolution using complex dense block on generative adversarial networks. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2866–2870. [Google Scholar]
  53. Fu, B.; Wang, L.; Wu, Y.; Wu, Y.; Fu, S.; Ren, Y. Weak texture information map guided image super-resolution with deep residual networks. Multimed. Tools Appl. 2022, 81, 34281–34294. [Google Scholar] [CrossRef]
  54. Cai, Q.; Li, J.; Li, H.; Yang, Y.H.; Wu, F.; Zhang, D. TDPN: Texture and Detail-Preserving Network for Single Image Super-Resolution. IEEE Trans. Image Process. 2022, 31, 2375–2389. [Google Scholar] [CrossRef]
  55. Meng, B.; Wang, L.; He, Z.; Jeon, G.; Dou, Q.; Yang, X. Gradient information distillation network for real-time single-image super-resolution. J.-Real-Time Image Process. 2021, 18, 333–344. [Google Scholar] [CrossRef]
  56. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  57. Agustsson, E.; Timofte, R. Ntire 2017 challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1122–1131. [Google Scholar]
  58. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Morel, M.L.A. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the Electronic Proceedings of the British Machine Vision Conference, Surrey, UK, 3–7 September 2012; pp. 1–10. [Google Scholar]
  59. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the International Conference on Curves and Surfaces, Avignon, France, 24–30 June 2010; pp. 711–730. [Google Scholar]
  60. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; pp. 416–423. [Google Scholar]
  61. Huang, J.B.; Singh, A.; Ahuja, N. Single image super resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  62. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  63. Fang, Z.; Zhao, M.; Yu, Z.; Li, M.; Yang, Y. A guiding teaching and dual adversarial learning framework for a single image dehazing. Vis. Comput. 2022, 38, 3563–3575. [Google Scholar] [CrossRef]
  64. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep laplacian pyramid networks for fast and accurate super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5835–5843. [Google Scholar]
  67. Wei, P.; Xie, Z.; Lu, H.; Zhan, Z.; Ye, Q.; Zuo, W.; Lin, L. Component divide-and-conquer for real-world image super-resolution. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 101–117. [Google Scholar]
  68. Ketsoi, V.; Raza, M.; Chen, H.; Yang, X. SREFBN: Enhanced feature block network for single-image super-resolution. IET Image Process. 2022, 16, 3143–3154. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed network.
Figure 1. The framework of the proposed network.
Mathematics 11 00216 g001
Figure 2. Dual-coordinate direction perception attention.
Figure 2. Dual-coordinate direction perception attention.
Mathematics 11 00216 g002
Figure 3. Cross-fusion block.
Figure 3. Cross-fusion block.
Mathematics 11 00216 g003
Figure 4. Training loss curve.
Figure 4. Training loss curve.
Mathematics 11 00216 g004
Figure 5. Visual comparisons with different methods for 2 × super-resolution on the Set5, Set14, B100, and Urban100 datasets.
Figure 5. Visual comparisons with different methods for 2 × super-resolution on the Set5, Set14, B100, and Urban100 datasets.
Mathematics 11 00216 g005
Figure 6. Visual comparisons with different methods for 3 × super-resolution on the Set5, Set14, B100, and Urban100 datasets.
Figure 6. Visual comparisons with different methods for 3 × super-resolution on the Set5, Set14, B100, and Urban100 datasets.
Mathematics 11 00216 g006
Figure 7. Visual comparisons with different methods for 4 × super-resolution on the Set5, Set14, B100, and Urban100 datasets.
Figure 7. Visual comparisons with different methods for 4 × super-resolution on the Set5, Set14, B100, and Urban100 datasets.
Mathematics 11 00216 g007
Figure 8. Comparison of ablation experiment results.
Figure 8. Comparison of ablation experiment results.
Mathematics 11 00216 g008
Figure 9. Hyper-parameter analysis on the DIV2K dataset.
Figure 9. Hyper-parameter analysis on the DIV2K dataset.
Mathematics 11 00216 g009
Table 1. Quantitative comparison of 2 × super-resolution results obtained by different methods.
Table 1. Quantitative comparison of 2 × super-resolution results obtained by different methods.
MethodScaleSet5Set14B100Urban100
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
SRCNN236.660.954232.450.906731.360.887929.500.8946
FSRCNN237.050.956032.660.909031.530.892029.880.9020
VDSR237.530.959033.050.913031.900.896030.770.9140
LapSRN237.520.959133.080.913031.080.895030.410.9101
CDC232.350.876629.200.793228.580.785626.060.7766
SREFBN237.920.959333.590.917532.170.899132.130.9279
Proposed238.070.971333.600.944232.200.903632.190.8549
Table 2. Quantitative comparison of 3 × super-resolution results obtained by different methods.
Table 2. Quantitative comparison of 3 × super-resolution results obtained by different methods.
MethodScaleSet5Set14B100Urban100
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
SRCNN332.750.909029.300.821528.410.786326.240.7989
FSRCNN333.180.914029.370.824028.530.791026.430.8080
VDSR333.670.921029.780.832028.830.799027.140.8290
LapSRN333.820.922729.870.832028.820.798027.070.8280
CDC325.290.296623.970.622424.740.616521.960.6087
SREFBN334.330.925730.280.840729.060.803828.030.8494
Proposed335.120.943230.370.855729.130.794128.270.7249
Table 3. Quantitative comparison of 4 × super-resolution results obtained by different methods.
Table 3. Quantitative comparison of 4 × super-resolution results obtained by different methods.
MethodScaleSet5Set14B100Urban100
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
SRCNN430.480.862827.500.751326.900.710124.520.7221
FSRCNN430.720.866027.610.755026.980.715024.620.7280
VDSR431.350.883028.020.768027.290.702625.180.7540
LapSRN431.540.885028.190.772027.320.727025.210.7560
CDC425.410.717724.010.616524.190.608521.500.5610
SREFBN432.010.891728.520.779027.500.733225.870.7787
Proposed432.270.903928.590.762427.600.707526.130.7893
Table 4. Ablation experiment performance.
Table 4. Ablation experiment performance.
MethodBaseTDFEMDCDPACFBPSNR/SSIM
Base 31.38/0.8333
Base + TDFEM 31.51/0.8456
Base + TDFEM + DCDPA 31.85/0.8509
Base + TDFEM + DCDPA + CFB32.19/0.8549
Table 5. Performance of the model with different numbers of GFBs.
Table 5. Performance of the model with different numbers of GFBs.
GFB Modulesn = 3n = 5n = 7n = 9
PSNR37.1038.0737.4037.24
SSIM0.96740.97130.96880.9676
Params (M)18.421.126.130.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Huang, Y.; Wang, K.; Qi, G.; Zhu, J. Single Image Super-Resolution Reconstruction with Preservation of Structure and Texture Details. Mathematics 2023, 11, 216. https://doi.org/10.3390/math11010216

AMA Style

Zhang Y, Huang Y, Wang K, Qi G, Zhu J. Single Image Super-Resolution Reconstruction with Preservation of Structure and Texture Details. Mathematics. 2023; 11(1):216. https://doi.org/10.3390/math11010216

Chicago/Turabian Style

Zhang, Yafei, Yuqing Huang, Kaizheng Wang, Guanqiu Qi, and Jinting Zhu. 2023. "Single Image Super-Resolution Reconstruction with Preservation of Structure and Texture Details" Mathematics 11, no. 1: 216. https://doi.org/10.3390/math11010216

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop