Next Article in Journal
Affirmative Ant Colony Optimization Based Support Vector Machine for Sentiment Classification
Next Article in Special Issue
Lightweight Path Recovery in IPv6 Internet-of-Things Systems
Previous Article in Journal
A Subspace Pre-Learning Strategy to Break the Interpose PUF
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Remote Sensing Image Super-Resolution Method and Its Application in Smart Cities

1
State Key Laboratory of Media Convergence and Communication, Communication University of China, Beijing 100024, China
2
Key Laboratory of Convergent Media and Intelligent Technology, Ministry of Education, Communication University of China, Beijing 100024, China
3
School of Computer and Cyber Sciences, Communication University of China, Beijing 100024, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(7), 1050; https://doi.org/10.3390/electronics11071050
Submission received: 11 March 2022 / Revised: 23 March 2022 / Accepted: 25 March 2022 / Published: 27 March 2022
(This article belongs to the Special Issue Edge Computing for Urban Internet of Things)

Abstract

:
With the growth of urban population, a series of urban problems have emerged, and how to speed up smart city construction has received extensive attention. Remote sensing images have the advantages of wide spatial coverage and rich information, and it is suitable for use as research data for smart cities. However, due to limitations in the imaging sensor conditions and complex weather, remote sensing images face the problems of insufficient resolution and cloud occlusion, which cannot meet the resolution requirements of smart city tasks. The remote sensing image super-resolution (SR) technique can improve the details and texture information without upgrading the imaging sensor system, which becomes a feasible solution for the above problems. In this paper, we propose a novel remote sensing image super-resolution method which leverages the texture features from internal and external references to help with SR reconstruction. We introduce the transformer attention mechanism to select and extract parts of texture features with high reference values to ensure that the network is lightweight, effective, and easier to deploy on edge computing devices. In addition, our network can automatically learn and adjust the alignment angles and scales of texture features for better SR results. Extensive comparison experiments show that our proposed method achieves superior performance compared with several state-of-the-art SR methods. In addition, we also evaluate the application value of our proposed SR method in urban region function recognition in smart cities. The dataset used in this task is low-quality. The comparative experiment between the original dataset and the SR dataset generated by our proposed SR method indicates that our method can effectively improve the recognition accuracy.

1. Introduction

In recent years, due to the continuous growth of urban population, a series of urban problems have emerged, such as traffic jams, excessive resource consumption, environmental pollution, and so on, which bring serious challenges to urban management and planning [1,2]. Smart city construction adopts intelligent technologies such as deep learning and the Internet of Things (IoT) to realize scientific and intelligent urban development, which has become an important target [3,4,5].
Remote sensing technology can capture urban land cover information from high altitudes, so that the generated remote sensing images have the advantages of wide spatial coverage and rich information, which have become important for urban big data and have been widely used in urban analysis, e.g., in urban heat island effect analysis [6], urban functional region layout [7,8], mapping and monitoring of slums in cities [9], and urban ecosystem analysis [10]. Recently, amid the increasing popularity of edge computing devices, which can collect or receive remote sensing images, such as unmanned aerial vehicles (UAVs) and vehicle-borne sensors, the application value of remote sensing images in smart cities has been further improved [11,12].
However, due to the limitations of imaging sensor conditions, weather, illumination, and so on, remote sensing images face the problems of insufficient resolution and cloud occlusion, which seriously affect its practical applicability in smart cities. The remote sensing image super-resolution (SR) technique can generate a clear and reasonable high-resolution (HR) image according to an input low-resolution (LR) image without upgrading the imaging sensor system. This has become a hot research topic in recent years.
Traditional SR approaches mainly rely on interpolation, including nearest neighbor interpolation, bilinear interpolation, and bicubic interpolation, but the performance of these approaches is poor. With the rapid development of deep learning technology, SR methods based on deep learning have attracted significant research attention in recent years.
Depending on whether the image degradation is known, existing SR methods can be divided into two categories: (1) non-blind SR methods, and (2) blind SR methods.
Non-blind means that the image has a fixed degradation type, such as Gaussian blur or bicubic down-sampling. In recent years, many methods have been proposed. Dong et al. [13] propose a pioneering method, SRCNN, which is aimed at bicubic degradation. The SRCNN method adopts a three-layer convolutional network to generate a realistic image. Then, an improved model based on SRCNN was proposed, named FSRCNN [14], greatly improved in speed and restoration quality. In addition, SRGAN [15] and ESRGAN [16] both adopt the idea of a generative adversarial network, and both improve the design of loss functions, and thus are able to yield more realistic results. For other degradation types, Zhang et al. [17] propose a dimensionality strategy, which explicitly takes the LR image and its degradation factors, i.e., blur kernel and noise level, as input, and can handle multiple and even spatially variant degradation. Considering that an image may have more than one degradation type and that the noise intensity of different degradation types is also different, Xu et al. [18] incorporate dynamic convolution, which is a far more flexible alternative for handling different degradations, with the result being superior performance. Other typical methods include EDSR [19], DRRN [20], RCAN [21], RDN [22], etc. These methods can only achieve good performance under specific degradation types. When the real degradation is different from the specific degradation, the performance will drop severely.
Another category of blind SR investigates how to reconstruct realistic results when the degradation type is unknown. The reason for image degradation is various, and it is difficult to predict in advance. Blind SR method is therefore more challenging. In recent years, many innovative blind SR methods have been proposed, which can be divided into two types. One is to estimate the degradation first, and then perform image reconstruction according to the estimated degradation type. Gu et al. [23] propose an iterative kernel correction method to gradually correct the blur kernel to approximate the real kernel as closely as possible, because only when the blur kernel matches can the SR results be realistic. Huang et al. [24] propose an iterative end-to-end trainable model combining a blur kernel estimator and an SR restorer, then repeat these two modules alternately, which can make the SR results more effective. The other is to directly perform image reconstruction without degradation estimation first. Wang et al. [25] propose an unsupervised degradation representation learning method which does not need to estimate the degradation explicitly, but instead learns abstract representations to distinguish various degradations in the representation space. They then introduce a degradation-aware SR (DASR) network to generate the SR results adaptively based on the learned degradation representations.
However, due to different image degradations and texture distributions, SR methods designed for natural images cannot obtain a good performance when applied to remote sensing images. At present, the SR method based on remote sensing images has attracted the attention of many researchers. For example, Lei et al. [26] proposed an SR method named local-global combined networks (LGCNet) to learn the multi-level representations between LR image and HR image. Haut et al. [27] proposed a new convolutional generator model to reconstruct the HR image without supervision. Xu et al. [28] proposed a deep memory connected network, which builds local and global memory connections to combine image detail with environmental information. Gu et al. [29] proposed a deep residual squeeze and excitation network. The network improved the feature representation effectively and reduced the parameters, and thereby achieved better accuracy and visual performance on two public remote sensing datasets. Wang et al. [30] proposed an adaptive multi-scale feature fusion network which can extract dense features from the original LR images directly without any interpolation operations. They then used several adaptive multi-scale feature extraction modules and adaptive gating mechanisms to extract and fuse the features.
However, these methods are usually estimated from input LR images. They always require a large amount of training samples and computing resources to efficiently learn rich prior knowledge, and are thus difficult to deploy on edge computing devices. Recently, reference-based approaches [31,32,33] have shown great potential in SR image synthesis. Most existing works implement SR in either internal reference or external reference, and few do so in both; the result is that they have limited prior information, leading to a lack of diversity and richness of image details.
In this paper, we leverage textural similarity and transferability to help in the restoration of remote sensing images. We propose a multi-scale texture transfer SR network (MSTT-SRNet) that combines internal reference and external reference to yield outstanding results. This reference information can provide rich and diverse prior knowledge, such as knowledge of colors, texture, and geometry, making it possible to generate realistic images. However, this reference information is redundant, and some even irrelevant. We introduce transformer attention mechanism [34] to effectively learn the similarity relationship between the LR texture features and Ref texture features so as to select the partly similar reference texture features and transfer them into LR images. This process can thus reduce the parameters and computation of the network and make it easier to deploy on edge computing devices. In addition, we design a multi-scale texture feature fusion module which can automatically learn and adjust the alignment angles and scales of the Ref texture feature for better results. In addition, we introduce a texture transfer loss to further enhance the effectiveness of multi-scale texture transfer.
In summary, the main contributions are as follows:
(1) We propose a novel remote sensing image SR method which leverages internal reference and external reference to jointly guide accurate restoration. Those references provide rich and diverse prior knowledge of colors, texture, and geometry, making it possible to generate realistic images. At the same time, we introduce the transformer attention mechanism to reduce the parameters and computation of the network, making it easier to deploy on edge computing devices.
(2) We design a multi-scale texture feature fusion module to automatically learn and adjust the alignment angles and scales of the Ref texture feature for better SR results.
(3) We conduct comparative experiments between our SR method and several other state-of-the-art SR methods. The comparative results show that our method achieves superior performance, which demonstrates the effectiveness of our method.
(4) In order to test the application effect of our proposed SR method in smart city tasks, we compare the accuracy improvement for urban region function recognition before and after super-resolution by our proposed method. The results show that the accuracy after super-resolution was significantly improved, which demonstrates that our SR method can effectively improve the performance of tasks based on remote sensing images in smart cities.
The rest of the paper is organized as follows: Section 2 explains the proposed SR method, including the overall network architecture, the details of two key modules, and the loss function. Section 3 explains the performance of the proposed method in the test dataset, and verifies the application effectiveness by applying our proposed method to the task of urban region function recognition. Section 4 provides the discussion and conclusion.

2. Method

In this section, we elaborate on how we devised the multi-scale texture transfer super-resolution network (MSTT-SRNet) to fully utilize both internal and external reference texture information for remote sensing image super-resolution. First, we briefly describe the overall network structure. Then, the similar texture feature extraction module (STFE) and multi-scale texture feature fusion module (MSTF2) are both elaborated in detail. Finally, the content of loss function is given.

2.1. Network Architecture

Figure 1 illustrates the overall network architecture of the proposed MSTT-SRNet. We deploy the well-known U-Net architecture as our backbone network. U-Net [35] is widely used in the fields of semantic segmentation and image super-resolution. It is composed of an encoder network (for semantic feature extraction) and a decoder network (for semantic information generation). The peer nodes between the encoder and the decoder are concatenated with each other to integrate semantic information and enhance the generated features. Compared to other backbone networks, U-Net has a relatively symmetrical structure which can well distinguish the low-resolution features from high-resolution features.
Unlike most SR methods, our network has two inputs and two outputs, as shown by the two data flows in Figure 1. The blue data flow indicates that the input is the LR image I l r and the output is the SR image I s r . The orange data flow indicates that the input is the Ref image I r e f and the output is the SR_Ref image I s r _ r e f .
In addition, our network combines internal reference and external reference. The input of internal reference is only the LR feature, and the input of external reference includes the LR feature and the Ref feature. These two references both involve two key modules: STFE and MSTF2. The STFE module calculates the similarity relationship between the LR texture feature and the Ref texture feature so as to extract the Ref texture features with high similarity and thus avoid the interference of irrelevant texture features, thereby reducing the complexity of the network. After similar texture feature extraction, the MSTF2 module was designed to transfer the aligned multi-scale Ref texture feature into the LR texture feature. We present the details of these two modules in what follows.

2.2. Similar Texture Feature Extraction Module

Intuitively, common objects from the same class usually share some similarity in texture and have a high similarity in features. To this end, we designed a similar texture feature extraction module to extract the Ref texture features, with high similarity and transferability, which contain the prior knowledge to generate SR results.
In this paper, the Ref texture features were mainly extracted from two aspects. (1) Internal reference: During the collection of remote sensing images, due to the problems of inappropriate angle, insufficient illumination, and cloud occlusion, the image content may be partially blurred; we can thus extract similar texture features from the image itself. As illustrated in Figure 2, the contents of the two blue boxes are very similar, but the content of the lower blue box is clearer, and thus the latter can be used as the prior knowledge for the content reconstruction of the upper blue box. (2) External reference: If we only rely on internal reference to reconstruct the LR image, the prior knowledge is low-quality and may not contain adequate texture details, limiting the performance in real-world applications. To this end, we leveraged a given high-resolution reference image to provide diverse and rich prior texture knowledge for super-resolution reconstruction, as shown by the two red boxes in Figure 2. The content of the red box in the reference image is very similar to the content of the red box in the low-resolution image, which can help with the content reconstruction.
The combination of internal reference and external reference provides rich and diverse prior knowledge. However, not all prior knowledge is useful for image reconstruction. Irrelevant and redundant texture features can even reduce the effectiveness of reference prior knowledge, making the network too complex to deploy on edge computing devices. The selection and extraction of texture features with high similarity from reference images is essential because accurate and proper texture features assist with the generation of SR images. However, it is challenging to extract such similar and transferable texture features for SR restoration process. Our key idea was to employ the transformer attention mechanism to learn the similarity relationship between LR texture features and Ref texture features, which can guide the extraction of similar texture features. Figure 3 illustrates the architecture of the proposed module.
Formally, we denote the LR texture feature as F l r R C × H × W and the Ref texture feature as F r e f R C × H × W , where C, H, and W represent the channel, height, and width of the feature maps, respectively. The transformer attention mechanism uses three learnable linear projection matrices, namely Query Q, Key K, and Value V, to compute the correlation relationship, and extracts the values based on the correlation relationship.
Specifically, in order to reduce the computational complexity and reduce redundant features, we first fed the input features F l r and F r e f into different convolution layers to generate feature maps of F l r Q , F r e f K , and F r e f V . The dimensions of these features were all reduced to C 2 × H × W .
Then, we performed reshape operations on F l r Q , F r e f K , and F r e f V , and converted the dimension into H W × C 2 , respectively. After this, we used these three feature maps to generate the linear projection matrices Q, K, and V to carry out similar texture feature extractions. The process can be formulated as follows:
Q = F l r Q × W Q ,
K = F r e f K × W K ,
V = F r e f V × W V ,
F s i m i l a r = s o f t m a x ( Q K T d k ) V ,
where W Q , W K , and W V are the learnable matrices, respectively. d k is the normalization factor.
Then, we performed reshape operations on F s i m i l a r , and converted the feature dimension to C 2 × H × W . Finally, we performed feature dimension reduction again to obtain the output Ref texture feature with high similarity, denoted by F o u t p u t ; the feature dimension was reduced to C 4 × H × W .
According to different references, in internal reference, F r e f is equal to F l r . In external reference, F r e f is the given reference texture feature. The similar texture feature extraction module enables our model to selectively extract more similar texture features and suppress irrelevant texture features, which can greatly enhance the effectiveness of reference texture features.

2.3. Multi-Scale Texture Feature Fusion Module

After obtaining highly similar Ref texture features, we considered transferring the Ref texture features into the LR texture features to improve the content quality. However, there were two important problems that needed to be solved: one was the texture alignment, the other the texture scale.
In this section, we propose a novel multi-scale texture feature fusion module for multi-scale Ref texture feature alignment and fusion. As shown in Figure 4, the proposed module adaptively adjusts the texture alignment angles and scales, which can work together to help generate better SR results.
Texture alignment. Since the objects in remote sensing images may appear in any direction and angle, the texture features between the LR image and the Ref image are likely to be misaligned. Therefore, we designed an alignment prediction module to predict the alignment angle for each convolution feature in the Ref features. Specifically, we denote the LR texture feature by F l r R C × H × W and the Ref texture feature by F r e f R C 4 × H × W , where C, H, and W represent the channel, height, and width of the feature maps, respectively. We first adopted a 3 × 3 convolution operation to encode the texture feature, and concatenated it with the LR feature. Then, a 1 × 1 × N prediction filter was employed to produce a feature map, with a size of N × H × W . N represents the alignment angles, which is 8 by default in our paper. Lastly, a softmax function was applied on the feature map to obtain the alignment angles for each feature matrix which will be convoluted in multi-scale texture feature extraction.
Texture scale. After obtaining the alignment angles, we extracted the multi-scale texture features from three scales. In the first type, we used a 1 × 1 convolution to extract the Ref texture feature if the scale was equal to the LR texture feature. In the second type, we used a 3 × 3 convolution with alignment angle to extract the Ref texture feature if the scale was larger than the LR texture feature. In the third type, we first × 2 up-sampled the Ref texture feature by bicubic interpolation, then used a 3 × 3 convolution with alignment angle to extract the Ref texture feature if the scale was smaller than the LR texture feature, and the stride of the convolution operation was set to 2. The convolution operation with alignment angle is described in the following.
A 3 × 3 convolution kernel can be expressed as K = { k 1 , k 2 , , k n } , where n represents the coordinate value in the convolution kernel; it is counted from left to right, then from top to bottom. The maximum value is 9. k i represents the value of the i t h coordinate point. F = { f 1 , f 2 , , f n } represents the corresponding feature matrix which will be convoluted with the 3 × 3 convolution kernel. A standard 2D convolution operation can be defined as follows:
F o u t p u t = ( F K ) ( n ) = i = 1 n ( f i k i )
In our study, we needed to rotate the texture feature F before the convolution operation according to the predicted alignment angle to re-sample a new texture feature matrix F ^ = { f 1 ^ , f 2 ^ , , f n } ^ . The re-sampling locations of different alignment angles are shown in Figure 5.
In this way, we were able to align the Ref texture feature to the LR texture feature. The new convolution operation can be defined as follows:
F o u t p u t ^ = ( F ^ K ) ( n ) = i = 1 n ( f i ^ k i )
Then, the aligned multi-scale Ref texture features were concatenated, and fused by a 1 × 1 convolution operation.
Finally, based on the aligned multi-scale Ref texture feature, a fusion operation was conducted to transfer the Ref texture feature into the LR texture feature, which was then used as the SR feature to generate the SR image.

2.4. Loss Function

In order to obtain better SR results, we used the weighted sum of multiple loss components to calculate the total loss function. The total loss function is defined as follows:
L t o t a l = λ c o n t L c o n t + λ p e r L p e r + λ t e x t L t e x t
where L c o n t denotes the content loss, L p e r denotes the perception loss, L t e x t denotes the texture transfer loss, and λ r e c , λ p e r , and λ t e x t are the balancing weights.
Content loss. The content loss consists of two parts: one is the content loss of the input LR image, and the other is the content loss of the Ref image. We used L 1 loss to measure the content loss, which was demonstrated better than L 2 loss and supported easier convergence. The calculation formula is defined as follows:
L c o n t = 1 C H W I s r I h r 1 + 1 C H W I s r r e f I r e f 1
where I s r and I h r denote the reconstructed image and the ground-truth high-resolution image of the input low-resolution image, respectively. I s r r e f and I r e f denote the reconstructed reference image and the input high-resolution reference image, respectively. C, H, and W represent the channel, height, and width of the images, respectively.
Perception loss. The content loss was too rigid to measure the content difference by pixel-to-pixel value. Perception loss measures the content difference through high-dimensional semantic features which are closer to the results of human perception. We measured the perception loss through the feature maps extracted by the VGG-19 network. The perception loss also includes two parts: one is the perception loss of the LR image, and the other is the perception loss of the Ref image. The formula is defined as follows:
L p e r = 1 C H W ϕ i ( I s r ) ϕ i ( I h r ) 1 + 1 C H W ϕ i ( I s r r e f ) ϕ i ( I r e f ) 1
where ϕ ( · ) denotes the VGG-19 network, and i represents the i-th layer’s feature map of the VGG-19 network. In this paper, we simply adopted the relu5_1 layer’s feature map. C, H, and W represent the channel, height, and width of the feature map, respectively.
Texture transfer loss. As the content loss and perception loss cannot guarantee that the texture of the SR image is transferred from the Ref image, we introduced a texture transfer loss to enhance the effectiveness of the texture transfer. With our texture transfer loss we attempted to restrict the SR texture feature so that it would be close to the Ref texture feature. We adopted the cosine similarity calculation to measure the texture similarity between the SR texture feature and the Ref texture feature. The texture transfer loss is defined as follows:
L t e x t = j = 0 3 ( 1 c o s ( F j s r , F j r e f ) )
where j denotes the number of the External+Internal module. c o s ( · ) represents the cosine function, the value range is normalized from 0 to 1, and the high value indicates the high similarity. F j s r denotes the output SR feature of external reference. F j r e f denotes the Ref texture feature which was fed into the external reference.

3. Experimental Results

In this section, we first describe the construction process of the dataset used in this paper and provide the implementation details. Then, we evaluate the performance of our method compared with other SR methods. Then, we fully verify the effectiveness of our designed modules by ablation experiments. Lastly, we test the application effect of our proposed method in smart cities by the task of urban region function recognition.

3.1. Datasets

Since there are no open available datasets for remote sensing images super-resolution with reference images, we chose the AID dataset [36] and WHU-RS19 dataset [37] to construct the training set and test set. These two datasets have relatively high pixel and spatial resolution.
The AID dataset was released by the Huazhong University of Science and Technology and Wuhan University in 2017. It contains 30 classes, with about 220–420 remote sensing images in each class. The total number is 10,000. The pixel resolution is 600 × 600 , and the spatial resolution is 0.5 m∼8 m.
The WHU-RS19 dataset was released by Wuhan University in 2011. It contains 19 classes. Each class has about 50 images with 600 × 600 pixels. The spatial resolution is 0.5 m. The total number is 1005.
Considering that the similarity between the LR image and the Ref image affects the super-resolution results significantly, we regarded the former image as the original HR image and the latter image as the corresponding Ref image in the same class. This strategy can ensure that the original HR image and the Ref image have a high texture similarity. In this way, we collected 5500 HR–Ref pair samples.
Figure 6 shows the image degradation process in our paper.
Finally, we obtained 22,000 pair samples in every scaling factor. Each pair sample included a low-resolution image, a reference image, and a high-resolution ground truth image. We selected 20,000 samples for training and 2000 samples for testing. Figure 7 shows some examples in the dataset.

3.2. Metrics

In this paper, we chose two standard evaluation metrics, peak signal-to-noise ration ( P S N R ) and the structural similarity index measure ( S S I M ) [38], as the objective measure indexes to evaluate the performance of our method and other compared SR methods.
We calculated the P S N R as follows:
M S E = 1 m n i = 0 m 1 j = 0 n 1 [ X ( i , j ) Y ( i , j ) ] 2 ,
P S N R = 10 · log 10 ( M A X X 2 M S E ) ,
where X denotes the HR image, Y denotes the SR image, m × n represents the pixel resolution of the image, and M A X denotes the maximum possible pixel value of the image, which was set to 255 in our experiments.
S S I M was calculated from three aspects: brightness, contrast, and structure. The value ranged from 0 to 1. The larger the value, the more similar the structure.
Brightness was calculated by mean as follows:
l ( X , Y ) = 2 μ X μ Y + c 1 μ X 2 + μ Y 2 + c 1 ,
Contrast was calculated by variance as follows:
c ( X , Y ) = 2 σ X σ Y + c 2 σ X 2 + σ Y 2 + c 2 ,
Structure was calculated by covariance as follows:
s ( X , Y ) = σ X Y + c 3 σ X σ Y + c 3 ,
where μ X and μ Y represent the mean value, σ X and σ Y represent the standard deviations, and σ X Y denotes the cross-covariance. c 1 = ( k 1 L ) 2 and c 2 = ( k 2 L ) 2 are two constants to avoid division by zero, where L is set to 255, and k 1 and k 2 are set to 0.01 and 0.03, respectively, in our experiments.
S S I M can be calculated as follows:
S S I M ( X , Y ) = [ l ( X , Y ) α · c ( X , Y ) β · s ( X , Y ) γ ] ,
Now, with α = 1 , β = 1 , γ = 1 , and c 3 = c 2 2 , the formula can be changed to:
S S I M ( X , Y ) = ( 2 μ X μ Y + c 1 ) ( 2 σ X Y + c 2 ) ( μ X 2 + μ Y 2 + c 1 ) ( σ X 2 + σ Y 2 + c 2 ) .

3.3. Implementation

While training the MSTT-SRNet, the batch size was set to 4. We used Adam [39] as the optimizer with a weight decay of 5 × 10 4 . We trained the network for 25 epochs. The initial earning rate was 0.001 and it decreased by a factor of 10 every 10 epochs. The multi-head in the transformer was set to 8. The weights of λ c o n t , λ p e r , and λ t e x t were 5.0, 1.0, and 1.0, respectively. Our proposed method was implemented by PyTorch, and all the experiments were run on a NVIDIA Titan XP GPU (12GB memory).

3.4. Comparisons with State-of-the-Art Methods

In order to better demonstrate the effectiveness of our method, we compared our method with several other SR methods. The compared methods were bicubic, FSRCNN [14], EDSR [19], TTSR [31], and DASR [25]. Bicubic is a typical traditional interpolation method. FSRCNN and EDSR are traditional non-blind super-resolution methods. TTSR is also a reference-based super-resolution method. DASR is a blind super-resolution method.
Figure 8 and Figure 9 show the visual reconstruction effect of scaling factors × 2 and × 4 , respectively. We can see that: (1) the images reconstructed by the bicubic method are very blurry, the detail information is missing, and the visual effectiveness is poor; (2) the images reconstructed by the FSRCNN method and the EDSR method demonstrate a great improvement in the content details, but there are still many cases where the boundary lines are blurred, and the reconstruction effect is not ideal; (3) The images reconstructed by the DASR method have a certain distortion, especially in the second group of samples, and color distortion occurs, the main reason being that the estimation of the image degradation type by this method does not deal with the degradation problem of cloud occlusion; (4) TTSR is also a reconstruction method based on texture reference, but the effect on remote sensing images is not ideal. The reason may be that the method is based on correlation analysis between low-resolution image features and the same down-sampled reference image features. However, there is no cloud occlusion degradation in the reference image, which affects the correlation learning between low-resolution features and reference features; (5) Our method learns the texture correlation between the reconstruct features and the reference features in the decoder stage, both of which have no cloud occlusion noise. Therefore, compared with other methods, the method proposed in this paper results in the best reconstruction effect, and the texture details are more abundant, which fully verifies the effectiveness of the method.
To further analyze the objective performance of our method, we reported the mean P S N R and S S I M values compared with other methods on test sets of various scaling factors. Table 1 shows the statistical results. It can be seen from the table that our method achieves better results than other state-of-the-art methods, which demonstrates the effectiveness of our method.

3.5. Analysis of Different Reference Images

In order to better analyze the reconstruction effectiveness as affected by different Ref images, Figure 10 shows the comparison of the generated SR results based on different Ref images. It can be seen from the figure that the texture similarity between the LR image and the Ref image on the left is low and that the SR result is poor, while the Ref image on the right has a high texture similarity with the LR image and the texture details of the corresponding SR result are richer, which fully indicates that it is effective for adopting a Ref image to help with SR reconstruction. At the same time, the Ref image with higher texture similarity has a higher reference value for helping with SR reconstruction.

3.6. Ablation Studies

In this section, we conduct several ablation experiments to verify the effectiveness of the key components in our method, including the similar texture feature extraction module, the multi-scale texture feature fusion module, and the texture transfer loss.
Similar texture feature extraction module. The function of the similar texture feature extraction module (STFE) lies in using a transformer attention mechanism to extract the high similarity texture feature from the Ref image, and in doing so to avoid the interference of irrelevant and redundant texture features to obtain more valuable reference prior knowledge.
To verify the effectiveness of the STFE module, we designed an ablation experiment. We removed the STFE module from the full model. For fairness, we also performed a 1 × 1 convolution operation on the Ref texture feature to reduce the feature dimension. Then, we fed the Ref texture feature to the MSTF2 module for the SR task.
Figure 11 shows the visual comparison results of scaling factor × 2 . We can see that both the LR image and the Ref image contain many objects, and the texture information is complex. Although they have a high texture similarity, the SR result from the network without STFE is poor, while the network with STFE restores richer texture details and a more realistic SR result, demonstrating the superior usefulness of the STFE module in the selection and extraction of Ref texture features.
Multi-scale texture feature fusion module. To study the effect of the MSTF2 module, we conducted an experiment by removing the MSTF2 module from the full model, then compared the SR results with those from the full model. In addition, we also conducted an experiment by cutting the playground area from the Ref image and then up-sampling by bicubic interpolation to 600 × 600 as the new Ref image, which means the different scales of the Ref images, though the experiment is not fully comparable. We show the performance comparison results of these two models based on differently scaled Ref images in Figure 12. As shown in the figure, under the same scale of the Ref image, the model with the MSTF2 module could obtain better SR results compared with the model without the MSTF2 module, which demonstrates the effectiveness of the proposed MSTF2 module. Under the different scales of Ref images, the results of the model without the MSTF2 module are superior when the scale of the Ref image is large, which implies that texture scale matters. By adding the MSTF2 module, the SR results of different scales are relatively approximate. This demonstrates the stronger adaptivity of multi-scale in extracting Ref texture features.
Texture transfer loss. To ensure that the texture transfer loss would benefit our network, we trained the MSTT-SRNet twice. The first training did not include texture transfer loss, while the second training did. Table 2 shows the quantitative comparison results of these two training networks with various scaling factors. We can see that: (1) in the scaling factor × 2 , texture transfer loss provides a 0.5 dB improvement in P S N R , and a 0.007 improvement in S S I M ; (2) in the scaling factor × 3 , texture transfer loss provides a 0.49 dB improvement in P S N R , and a 0.009 improvement in S S I M ; (3) in the scaling factor × 4 , texture transfer loss provides a 0.34 dB improvement in P S N R , and a 0.012 improvement in S S I M . The above results show the effectiveness of texture transfer loss.

3.7. Application in Smart Cities

In this section, we evaluate the application effect of our proposed SR method in smart cities. We applied our SR method to a challenging large-scale dataset for urban region function recognition (URFR dataset). The dataset was released by the 5th Baidu and XJTU Big Data Contest in 2019. The aim of the contest was to design an urban region function classification model based on remote sensing images and user visit data. The dataset consists of two parts: the first part was provided in the preliminary round, denoted as URFR-1, and includes 40,000 training samples and 10,000 test samples. The second part was provided in the semi-final round, denoted as URFR-2, and included 400,000 training samples and 100,000 test samples. There were nine types of region function labels in the dataset, namely residential area (res.), school (sch.), industrial park (ind.), railway station (rail.), airport (air.), park (park), shopping area (shop.), administrative district (adm.), and hospital (hosp.). In our experiments, we only used the remote sensing images in the training samples with ground truth labels. This dataset was particularly challenging as the resolution of the remote sensing images was 100 × 100 , and the image content was blurry. Some images even had serious cloud occlusion, which led to poor performance based on this dataset. Table 3 shows the details of the URFR-1 and URFR-2 datasets.
Visual Comparison. Figure 13 shows some remote sensing images in the URFR dataset and the corresponding SR results generated by our proposed method. From the visual comparisons, we can see that the SR results reconstructed by our proposed method have clearer edges and richer textures.
Quantitative Evaluation. We performed the urban region function recognition experiments on three widely used convolutional neural networks, namely VGG-19 [40], ResNet-50 [41], and DenseNet-121 [42], to evaluate the application effect. We conducted a comparative experiment. (1) Before SR: We used the original URFR dataset to train in turn the three classification networks pre-trained on ImageNet. (2) After SR: We applied our proposed SR method to the URFR dataset to generate the SR images. Then, we used the SR URFR dataset to train in turn the three classification networks pre-trained on ImageNet again. In our experiment, we set the same training parameters to the same type of classification network during the two comparative training processes. We adopted the SR results of scaling factor × 3 . We resized both the original images and the SR images to 224 × 224 for training and testing.
Table 4 summarizes the accuracy results of the comparative experiments on the URFR-1 and URFR-2 test sets. From the results, we can see that the accuracy results after SR of the three classification networks all outperformed the results before SR. The accuracy of VGG-19, ResNet-50, and DenseNet-121 on the URFR-1 test set were improved by 2.32%, 4.11%, and 4.04%, respectively. The accuracy of VGG-19, ResNet-50, and DenseNet-121 on the URFR-2 test set were improved by 2.52%, 3.78%, and 4.41%, respectively. This indicates that our proposed SR method can effectively improve remote sensing image quality and enhance the application value in smart-city-related tasks.

4. Discussion and Conclusions

In this paper, to address the problems of insufficient resolution and blurry content of remote sensing images, we proposed a novel SR method, called the multi-scale texture transfer super-resolution network (MSTT-SRNet), which effectively leverages internal and external reference prior knowledge. Particularly, the proposed method introduces the transformer attention mechanism in order to exploit the texture similarity measurement between the LR image and the Ref image to enhance the transfer texture feature representation and reduce the parameters and computation of the network. At the same time, considering the diverse scales and alignment angles of texture features, a multi-scale texture feature fusion module was designed to obtain more accurate reconstruction results. The experimental results show that our proposed method can achieve better subjective and objective performance compared with several state-of-the-art SR methods.
In addition, we also evaluated the application value of our proposed SR method in smart cities through the task of urban region function recognition. The dataset used in this task was particularly challenging as the resolution of the remote sensing images was very low ( 100 × 100 ) and the image content was blurry; some images even had serious cloud occlusion, which led to poor performance. We applied our proposed SR method to this low-quality dataset. We then conducted comparative experiments between the SR dataset and the original dataset on three widely used convolutional neural networks: VGG-19, ResNet-50, and DenseNet-121. The results show that our proposed SR method can effectively increase the accuracy of urban region function recognition. This indicates that our proposed SR method has high application value in smart cities.
However, we found some problems during the experiment. Compared with real-world datasets, the synthetic dataset in our paper based on the AID and WHU-RS19 datasets did not contain sufficient variant scales. Thus, the texture scale difference between LR and Ref images was small, which led to inadequate learning in the extraction and fusion of multi-scale texture features. In the future, we will explore the variant scales in the process of dataset construction. The possible ways to do this include choosing source datasets with different spatial resolutions and selecting LR–Ref pair samples with scale variation. Moreover, we also need to further test the application effect of our proposed method in more smart city tasks based on remote sensing images.

Author Contributions

Conceptualization, N.Z. and Y.W.; methodology, N.Z., Y.W. and S.F.; validation, N.Z. and S.F.; formal analysis, Y.W. and S.F.; investigation, N.Z.; data curation, N.Z.; writing—original draft preparation, N.Z.; writing—review and editing, Y.W. and S.F.; visualization, N.Z.; supervision, Y.W. and S.F.; project administration, Y.W. and S.F.; funding acquisition, S.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China under Grant 2019YFB1406204.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zheng, Y.; Capra, L.; Wolfson, O.; Yang, H. Urban computing: Concepts, methodologies, and applications. ACM Trans. Intell. Syst. Technol. (TIST) 2014, 5, 38. [Google Scholar] [CrossRef]
  2. Dameri, R.P. Searching for smart city definition: A comprehensive proposal. Int. J. Comput. Technol. 2013, 11, 2544–2551. [Google Scholar] [CrossRef]
  3. Kim, T.-h.; Ramos, C.; Mohammed, S. Smart City and Iot. In Future Generation Computer Systems; Elsevier: Amsterdam, The Netherlands, 2017; Volume 76, pp. 159–162. [Google Scholar]
  4. Kirimtat, A.; Krejcar, O.; Kertesz, A.; Tasgetiren, M.F. Future trends and current state of smart city concepts: A survey. IEEE Access 2020, 8, 86448–86467. [Google Scholar] [CrossRef]
  5. Hashem, I.A.T.; Chang, V.; Anuar, N.B.; Adewole, K.; Yaqoob, I.; Gani, A.; Ahmed, E.; Chiroma, H. The role of big data in smart city. Int. J. Inf. Manag. 2016, 36, 748–758. [Google Scholar] [CrossRef] [Green Version]
  6. Zhou, D.; Xiao, J.; Bonafoni, S.; Berger, C.; Deilami, K.; Zhou, Y.; Frolking, S.; Yao, R.; Qiao, Z.; Sobrino, J.A. Satellite remote sensing of surface urban heat islands: Progress, challenges, and perspectives. Remote Sens. 2019, 11, 48. [Google Scholar] [CrossRef] [Green Version]
  7. Herold, M.; Scepan, J.; Clarke, K.C. The use of remote sensing and landscape metrics to describe structures and changes in urban land uses. Environ. Plan. A 2002, 34, 1443–1458. [Google Scholar] [CrossRef] [Green Version]
  8. Xu, L.; Jing, W.; Song, H.; Chen, G. High-resolution remote sensing image change detection combined with pixel-level and object-level. IEEE Access 2019, 7, 78909–78918. [Google Scholar] [CrossRef]
  9. Mahabir, R.; Croitoru, A.; Crooks, A.T.; Agouris, P.; Stefanidis, A. A critical review of high and very high-resolution remote sensing approaches for detecting and mapping slums: Trends, challenges and emerging opportunities. Urban Sci. 2018, 2, 8. [Google Scholar] [CrossRef] [Green Version]
  10. Tavares, P.A.; Beltrão, N.; Guimarães, U.S.; Teodoro, A.; Gonçalves, P. Urban ecosystem services quantification through remote sensing approach: A systematic review. Environments 2019, 6, 51. [Google Scholar] [CrossRef] [Green Version]
  11. Tenedório, J.A.; Estanqueiro, R.; Matos Lima, A.; Marques, J. Remote sensing from unmanned aerial vehicles for 3d urban modelling: Case study of loulé, portugal. In Back to the Sense of the City: International Monograph Book; Centre of Land Policy and Valuations (CPSV): Krakow, Poland, 2016; pp. 1118–1128. [Google Scholar]
  12. Wang, Y.; Cheng, L.; Chen, Y.; Wu, Y.; Li, M. Building point detection from vehicle-borne lidar data based on voxel group and horizontal hollow analysis. Remote Sens. 2016, 8, 419. [Google Scholar] [CrossRef] [Green Version]
  13. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 184–199. [Google Scholar]
  14. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 391–407. [Google Scholar]
  15. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  16. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Loy, C.C. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  17. Zhang, K.; Zuo, W.; Zhang, L. Learning a single convolutional super-resolution network for multiple degradations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3262–3271. [Google Scholar]
  18. Xu, Y.-S.; Tseng, S.-Y.R.; Tseng, Y.; Kuo, H.-K.; Tsai, Y.-M. Unified dynamic convolutional network for super-resolution with variational degradations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12496–12505. [Google Scholar]
  19. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  20. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3147–3155. [Google Scholar]
  21. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  22. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  23. Gu, J.; Lu, H.; Zuo, W.; Dong, C. Blind super-resolution with iterative kernel correction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1604–1613. [Google Scholar]
  24. Huang, Y.; Li, S.; Wang, L.; Tan, T. Unfolding the alternating optimization for blind super resolution. Adv. Neural Inf. Process. Syst. 2020, 33, 5632–5643. [Google Scholar]
  25. Wang, L.; Wang, Y.; Dong, X.; Xu, Q.; Yang, J.; An, W.; Guo, Y. Unsupervised degradation representation learning for blind super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10581–10590. [Google Scholar]
  26. Lei, S.; Shi, Z.; Zou, Z. Super-resolution for remote sensing images via local–global combined network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1243–1247. [Google Scholar] [CrossRef]
  27. Haut, J.M.; Fernandez-Beltran, R.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Pla, F. A new deep generative network for unsupervised remote sensing single-image super-resolution. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6792–6810. [Google Scholar] [CrossRef]
  28. Xu, W.; Guangluan, X.; Wang, Y.; Sun, X.; Lin, D.; Yirong, W. High quality remote sensing image super-resolution using deep memory connected network. In Proceedings of the IGARSS 2018—2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 8889–8892. [Google Scholar]
  29. Gu, J.; Sun, X.; Zhang, Y.; Fu, K.; Wang, L. Deep residual squeeze and excitation network for remote sensing image super-resolution. Remote Sens. 2019, 11, 1817. [Google Scholar] [CrossRef] [Green Version]
  30. Wang, X.; Wu, Y.; Ming, Y.; Lv, H. Remote sensing imagery super resolution based on adaptive multi-scale feature fusion network. Sensors 2020, 20, 1142. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Yang, F.; Yang, H.; Fu, J.; Lu, H.; Guo, B. Learning texture transformer network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 5791–5800. [Google Scholar]
  32. Zhang, Z.; Wang, Z.; Lin, Z.; Qi, H. Image super-resolution by neural texture transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7982–7991. [Google Scholar]
  33. Huang, J.-B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  34. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. [Google Scholar]
  35. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  36. Xia, G.S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L. Aid: A Benchmark Dataset for Performance Evaluation of Aerial Scene Classification. IEEE Trans. Geosci. Remote Sens. 2016, 55, 3965–3981. [Google Scholar] [CrossRef] [Green Version]
  37. Sheng, G.; Yang, W.; Xu, T.; Sun, H. High-resolution satellite scene classification using a sparse coding based multiple feature combination. Int. J. Remote Sens. 2012, 33, 2395–2412. [Google Scholar] [CrossRef]
  38. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  39. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  40. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 21–26 July 2016; pp. 770–778. [Google Scholar]
  42. Huang, G.; Liu, Z.; Maaten, L.V.D.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
Figure 1. The structure of the proposed MSTT-SRNet: the input LR image flow is represented by the blue lines, and the reference image flow by the orange lines. STFE denotes the similar texture feature extraction module. MSTF2 denotes the multi-scale texture feature fusion module.
Figure 1. The structure of the proposed MSTT-SRNet: the input LR image flow is represented by the blue lines, and the reference image flow by the orange lines. STFE denotes the similar texture feature extraction module. MSTF2 denotes the multi-scale texture feature fusion module.
Electronics 11 01050 g001
Figure 2. Examples of internal reference and external reference.
Figure 2. Examples of internal reference and external reference.
Electronics 11 01050 g002
Figure 3. Illustration of the similar texture feature extraction module (STFE). The feature maps F o u t p u t are fed into the MSTF2 module as the input feature maps F r e f .
Figure 3. Illustration of the similar texture feature extraction module (STFE). The feature maps F o u t p u t are fed into the MSTF2 module as the input feature maps F r e f .
Electronics 11 01050 g003
Figure 4. The structure of the proposed multi-scale texture feature fusion module (MSTF2).
Figure 4. The structure of the proposed multi-scale texture feature fusion module (MSTF2).
Electronics 11 01050 g004
Figure 5. The re-sampling locations of different alignment angles. The total number of alignment angles is 8.
Figure 5. The re-sampling locations of different alignment angles. The total number of alignment angles is 8.
Electronics 11 01050 g005
Figure 6. The procedures of image degradation. Gaussian kernel size was randomly sampled from { 5 , 7 , 9 , 11 } . The cloud brightness was randomly sampled from {220:240}. The cloud density was randomly sampled from { 0.01 , 0.02 , 0.03 , 0.04 , 0.05 } .
Figure 6. The procedures of image degradation. Gaussian kernel size was randomly sampled from { 5 , 7 , 9 , 11 } . The cloud brightness was randomly sampled from {220:240}. The cloud density was randomly sampled from { 0.01 , 0.02 , 0.03 , 0.04 , 0.05 } .
Electronics 11 01050 g006
Figure 7. Examples from the training dataset.
Figure 7. Examples from the training dataset.
Electronics 11 01050 g007
Figure 8. Visual comparison between our method and other methods with a × 2 scaling factor.
Figure 8. Visual comparison between our method and other methods with a × 2 scaling factor.
Electronics 11 01050 g008
Figure 9. Visual comparison between our method and other methods with a × 4 scaling factor.
Figure 9. Visual comparison between our method and other methods with a × 4 scaling factor.
Electronics 11 01050 g009
Figure 10. SR results using different reference images with a × 2 scaling factor.
Figure 10. SR results using different reference images with a × 2 scaling factor.
Electronics 11 01050 g010
Figure 11. Compared SR results between MSTT-SRNet without and with STFE module based on the same Ref image with a × 2 scaling factor.
Figure 11. Compared SR results between MSTT-SRNet without and with STFE module based on the same Ref image with a × 2 scaling factor.
Electronics 11 01050 g011
Figure 12. Compared SR results between MSTT-SRNet without and with the MSTF2 module based on different scale Ref images with a × 2 scaling factor.
Figure 12. Compared SR results between MSTT-SRNet without and with the MSTF2 module based on different scale Ref images with a × 2 scaling factor.
Electronics 11 01050 g012
Figure 13. Visual comparisons on URFR dataset with a × 3 scaling factor.
Figure 13. Visual comparisons on URFR dataset with a × 3 scaling factor.
Electronics 11 01050 g013
Table 1. The mean P S N R (dB) and S S I M on a test set with scale factors of × 2 , × 3 , and × 4 .
Table 1. The mean P S N R (dB) and S S I M on a test set with scale factors of × 2 , × 3 , and × 4 .
ScaleBicubicFSRCNNEDSRTTSRDASROurs
P S N R /dB × 2 29.7131.6431.8631.9430.8332.31
× 3 27.3928.2128.5529.0328.8729.86
× 4 25.4226.9127.0127.2226.3327.65
S S I M × 2 0.8950.9020.9040.9130.8990.918
× 3 0.8120.8450.8410.8560.8540.866
× 4 0.7810.8030.8110.8120.8060.831
Table 2. The compared mean P S N R (dB) and S S I M on a test set trained by MSTT-SRNet without and with texture transfer loss with scale factors of × 2 , × 3 , and × 4 .
Table 2. The compared mean P S N R (dB) and S S I M on a test set trained by MSTT-SRNet without and with texture transfer loss with scale factors of × 2 , × 3 , and × 4 .
Scaling FactorTexture Transfer Loss PSNR /dB SSIM
× 2 without31.810.911
with32.310.918
× 3 without29.370.857
with29.860.866
× 4 without27.310.819
with27.650.831
Table 3. The split settings and per-class numbers of the URFR-1 and URFR-2 datasets.
Table 3. The split settings and per-class numbers of the URFR-1 and URFR-2 datasets.
DatasetSplitres.sch.ind.rail.air.parkshop.adm.hosp.Total
URFR-1train81526417303111552947468529732209243134,000
val2171889840851408768771000
test11739334611634326824573403595000
URFR-2train102,21177,52043,315559214,00053,30017,98711,19114,884340,000
val309422011338170410155851432139410,000
test15,06511,33263628262084782626341669220250,000
Table 4. The accuracy results of the comparative experiments on the URFR-1 and URFR-2 test sets.
Table 4. The accuracy results of the comparative experiments on the URFR-1 and URFR-2 test sets.
NetworkImage TypeAccuracy (URFR-1 %)Accuracy (URFR-2 %)
VGG-19original images39.8153.82
SR images42.1356.34
ResNet-50original images44.4854.72
SR images48.5958.50
DenseNet-121original images45.0855.43
SR images49.1259.84
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, N.; Wang, Y.; Feng, S. A Lightweight Remote Sensing Image Super-Resolution Method and Its Application in Smart Cities. Electronics 2022, 11, 1050. https://doi.org/10.3390/electronics11071050

AMA Style

Zhang N, Wang Y, Feng S. A Lightweight Remote Sensing Image Super-Resolution Method and Its Application in Smart Cities. Electronics. 2022; 11(7):1050. https://doi.org/10.3390/electronics11071050

Chicago/Turabian Style

Zhang, Nenghuan, Yongbin Wang, and Shuang Feng. 2022. "A Lightweight Remote Sensing Image Super-Resolution Method and Its Application in Smart Cities" Electronics 11, no. 7: 1050. https://doi.org/10.3390/electronics11071050

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop