Next Article in Journal
Assessment of NADH/NAD+ Redox Imbalance in Psoriatic Lesions Using the FMSF Technique: Therapeutic Aspects
Next Article in Special Issue
A Survey of Deep Learning Road Extraction Algorithms Using High-Resolution Remote Sensing Images
Previous Article in Journal
Overcoming Periodic Stripe Noise in Infrared Linear Array Images: The Fourier-Assisted Correlative Denoising Method
Previous Article in Special Issue
Scale-Hybrid Group Distillation with Knowledge Disentangling for Continual Semantic Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Thermal Image Super-Resolution Based on Lightweight Dynamic Attention Network for Infrared Sensors

School of Mechatronic Engineering and Automation, Shanghai University, Shanghai 200444, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8717; https://doi.org/10.3390/s23218717
Submission received: 22 September 2023 / Revised: 22 October 2023 / Accepted: 23 October 2023 / Published: 25 October 2023

Abstract

:
Infrared sensors capture infrared rays radiated by objects to form thermal images. They have a steady ability to penetrate smoke and fog, and are widely used in security monitoring, military, etc. However, civilian infrared detectors with lower resolution cannot compare with megapixel RGB camera sensors. In this paper, we propose a dynamic attention mechanism-based thermal image super-resolution network for infrared sensors. Specifically, the dynamic attention modules adaptively reweight the outputs of the attention and non-attention branches according to features at different depths of the network. The attention branch, which consists of channel- and pixel-wise attention blocks, is responsible for extracting the most informative features, while the non-attention branch is adopted as a supplement to extract the remaining ignored features. The dynamic weights block operates with 1D convolution instead of the full multi-layer perceptron on the global average pooled features, reducing parameters and enhancing information interaction between channels, and the same structure is adopted in the channel attention block. Qualitative and quantitative results on three testing datasets demonstrate that the proposed network can superior restore high-frequency details while improving the resolution of thermal images. And the lightweight structure of the proposed network with lower computing cost can be practically deployed on edge devices, effectively improving the imaging perception quality of infrared sensors.

1. Introduction

RGB sensors are universally used in smartphones, drones, laptops, and other devices due to their excellent imaging quality and speed. However, the image quality captured by RGB sensors degrades dramatically under harsh conditions [1]. The wavelength of long-wave infrared radiation ranges between 7 and 14 μm, while the wavelength of visible light in the electromagnetic spectrum lies between 390 and 780 nm. Therefore, infrared sensors have a robust ability to penetrate smoke, fog, and haze, and can replace RGB sensors in the aforementioned scenarios. Nonetheless, high-resolution (HR) infrared focal plane arrays are expensive, and civilian uncooled infrared detectors output low-resolution (LR) thermal images that cannot match the megapixel RGB sensors [2]. The straightforward approach is to design complex hardware devices to enhance the resolution of thermal images, but the extended manufacturing time limits their practical application. In recent years, a method known as super-resolution (SR) has been extensively developed to enhance LR image resolution [3].
The purpose of image super-resolution (ISR) is to recover the corresponding HR images from the observed degraded LR counterparts. The current ISR methods can be classified into three categories based on technical approaches: interpolation based, reconstruction based, and learning based. The linear interpolation algorithms (e.g., nearest, bilinear, and bicubic) design interpolation weighting functions based on the assumption of the local smoothness of the image [4]. They are simple and fast but are prone to aliasing artifacts in areas with rich high-frequency information. To overcome the shortcomings of linear interpolation, more works focus on nonlinear interpolation algorithms, i.e., adaptive image interpolation. For example, adaptive interpolation methods based on edge guidance [5,6] or perception [7] and non-local means [8] improve the perceptual quality of interpolated images, but high-frequency details are still insufficient under large scale factors. The ISR methods based on pixel domain reconstruction employ the introduced prior knowledge as a constraint to iteratively solve the objective function until it converges to the local optimal solution. Projection onto convex sets, iterative back projection and maximum a posteriori (MAP) are representative algorithms [4]. The reconstruction-based ISR methods [4,9,10] first establish a degradation model from HR images to LR images, which is the inverse problem of ISR [11,12,13]. Therefore, the degradation model can be reversely solved based on algorithms such as MAP estimation, and the HR image can be predicted. For instance, Greenbaum et al. [13] generated HR images based on super-resolved stacks of multiple shifted LR images to improve the field-of-view of dense samples captured by lensfree holographic microscopy. As a shallow learning algorithm based on learning, traditional sparse dictionary-based ISR methods suffer from slow speed and are limited by the size of the overcomplete dictionary, resulting in unsatisfying performance [14]. However, with the development of multimedia technology, the above methods have gradually reached performance bottlenecks and cannot meet the needs of generating high-definition or even ultra-high-definition images. Recently, ISR models based on deep convolutional neural networks (CNNs) achieved impressive performance gains over traditional methods [1,3]. CNN-based ISR methods map LR images to HR images in an end-to-end manner according to the datasets during training. Taking advantage of the powerful nonlinear fitting and automatic feature learning capabilities of CNNs, as well as the emergence of dedicated acceleration hardware such as the neural processing unit (NPU), the performance of the ISR network trained with massive training data is significantly better than the above traditional methods. SRCNN [15,16], as the first CNN-based ISR model, exhibits significantly superior performance than interpolation- and sparse coding-based methods. Relying on the powerful nonlinear fitting ability of CNN, CNN-based ISR models for RGB sensors have been continuously proposed [17]. However, there are relatively few ISR methods specifically designed for infrared sensors. Therefore, there is an urgent need to develop a thermal image SR model based on CNN that can be applied in practice.
Inspired by the human visual system (HVS), progressively ISR networks employ attention mechanisms to improve performance [18]. The HVS is not equally capable of processing all the information contained in the observed scene but concentrates limited computing resources on the most information-rich regions. For instance, when humans gaze at the sky, their attention is likely to be focused on objects like birds flying in the sky rather than the background of the sky itself. Similarly, it is the high-frequency details such as edges and textures rather than smooth areas that most affect the perceived quality of an image. The purpose of ISR is to maximize the recovery of high-frequency information while improving the resolution. Channel attention (CA) and spatial attention (SA) mechanisms are currently widely used attention methods and have been proven to improve many low-level computer vision tasks, including ISR [19,20,21]. CA and SA perform recalibration operations in the channel and pixel spaces of feature maps, respectively. To make full use of channel and spatial information interaction, many works stack CA and SA blocks to form attention modules and reuse them [1,22,23]. Previous studies [18,24] have shown that although LR images have insufficient resolution, they still contain a large amount of low-frequency information and high-frequency details. ISR networks that without attention blocks process all channels and image regions equally cannot effectively recover high-frequency details. Therefore, the attention mechanism can enrich the edges and texture details of the final super-solved HR image, thereby improving the perceptual quality.
The few existing CNN-based thermal image SR models for infrared sensors mainly suffer from the following three shortcomings. (1) Inefficient sequential stacking of attention modules. As shown by Chen et al. [18], simply using the same attention module hierarchy to extract features is not always beneficial to the final RGB ISR performance. And we prove in Section 3 that thermal image SR for infrared sensors should embrace attention modules variously at different depths of the network. The results show that the early attention feature extraction modules enhance low-level information, the tail modules extract high-frequency details, and the middle attention modules enhance the above two features mixedly. Therefore, it is necessary to dynamically adjust the attention weights according to the characteristics of different stages of the network. (2) Networks are complex and unwieldy. With the advent of residual structures [25], it is possible to stably train very deep networks. Since then, more and more ISR networks have improved performance by continuously increasing network capacity (i.e., increasing the number of layers of the network and the width of each layer). For example, EDSR [26] has more than 40M parameters, and the huge computing power requirement makes it almost impossible to deploy in embedded devices, which limits its practical application. Although well-designed networks with more parameters can continuously improve performance, the resulting gains and training/inference costs have to be considered. In other words, there should be a trade-off between the performance and scope of parameters. (3) The designed attention module is not compact. Inspired by SENet [27], more researchers [28,29,30] have invested in designing complex CA structures to enhance channel dimension features extraction, or combining convoluted SA blocks to improve performance. Although these methods enrich the details of the super-solved images, the heavy computational burden leads to slow inference, limiting their practical applications. Our initial idea was to design an efficient attention module that could reduce the complexity of the network and achieve satisfactory performance, making it possible to deploy on edge devices.
To address the above issues, we propose a lightweight dynamic attention super-resolution network (LDASRNet) for infrared sensors to super-resolve LR thermal images. The LDASRNet consists of a shallow feature extraction (SFE) module, a deep feature extraction (DFE) module and a feature reconstruction (FRec) module. The SFE module consists of only one convolutional layer with filter size 3 × 3, which is used to extract shallow features. The DFE module consists of sequentially stacked dynamic attention blocks (DABs). Different from previous works, our proposed DAB adaptively and dynamically assigns weights to attention and non-attention branches according to different deep features of the network. In particular, the efficient CA block efficiently aggregates channel dimension features. And the pixel attention block generates 3D instead of 2D attention maps compared with SA, and obtains more performance gains with less computational cost. The FRec module is used to reconstruct the final HR thermal image.
In this paper, our main contributions are as follows:
  • Aiming at the problem of the low perception quality of civilian LR uncooled infrared sensors, we propose a CNN-based dynamic attention network to super-resolve LR thermal images. The proposed network outperforms compared models with more parameters while maintaining a lightweight structure, showing the potential to be deployed on a handheld thermal imaging camera devices with limited computing power. We train the proposed network with a mixture of data augmentation methods, and experiments show that multiple pixel domain data augmentation methods can effectively improve the ISR performance.
  • We propose a lightweight DAB. DAB at different stages dynamically reweights the attention and non-attention branches according to the input feature maps. Furthermore, our proposed attention branch consists of an efficient CA block and a residual pixel attention block; the latter differs from existing SA mechanisms that only generate 2D attention maps but use 3D attention maps to enhance spatial features.
  • Qualitative and quantitative results on multiple datasets demonstrate that the proposed network can effectively improve thermal image resolution while recovering visually pleasing high-frequency details. Specifically, the proposed network achieves the highest performance metrics on the public testing dataset with ×3 scale factor. And compared to the second-best model A 2 N , the scope of parameters is only 34% of latter (i.e., 0.34 M vs. 1.0 M).
The remainder of this paper is organized as follows. Section 2 describes the technical details of the existing attention mechanism, as well as CNN-based SR methods for RGB and thermal images. Section 3 shows the motivation and necessity of our proposed dynamic attention for thermal image SR. Section 4 details our proposed dynamic attention network structure. Section 5 qualitatively and quantitatively compares the proposed network with state-of-the-art lightweight ISR models. We conclude this paper in Section 6.

2. Related Work

2.1. CNN-Based Image Super-Resolution

Traditional ISR methods based on the sparse coding framework represent HR image patches as sparse linear combinations of atoms in an over-complete dictionary [14]. Therefore, the ISR performance is limited by the dictionary size, and the inference speed is slow, resulting in dissatisfactory ISR performance under large scale factors [1].
As CNNs have demonstrated impressive accuracy in the field of image recognition [31], CNN-based ISR methods have emerged. SRCNN [15], as a pioneering CNN-based ISR work, achieves significantly superior performance than traditional ISR methods. And FSRCNN [16] further boosts SRCNN to obtain more gains at a lower computational cost. However, SRCNN only has three convolutional layers, resulting in a relatively small model capacity, which may not be able to recover sufficient details when faced with large-scale upsampling scale factors. Following SRCNN, VDSR [32] explores multi-layer small-size convolution kernels to expand the receptive field, and the resulting 20-layer network greatly improves accuracy. As the residual connection in ResNet [25] makes it possible to stably train very deep networks, the residual structure is widely used in the ISR models. Unlike SRResNet [33], EDSR [26] consists of 32 residual blocks that remove batch normalization, reducing memory consumption and artifacts. However, the above model adopts the pre-upsampling method, that is, the LR input is first interpolated to the desired size, which increases the difficulty and computational burden of subsequent deep feature extraction. Therefore, more works adopt a post-upsampling structure, which upsample features to the expected resolution at the end of the model [3]. RDN [34] employs residual blocks with dense connections to make full use of the features of previous layers, and the formed persistent memory mechanism effectively utilizes abundant features. Inspired by SENet [27], RCAN [24] models interactions between channels to recalibrate the channel features. As a scale-attention-based network, RCAN adopts a two-level residual CA structure to deepen the network depth while utilizing more low-frequency information. Similarly, HAN [35] with channel–spatial attention modules to explore the relationship between pixel domain and channel dimension, and uses layer attention to utilize the output features of all residual groups. The effectiveness of CA and SA to reweight features in channel and spatial dimensions respectively to enhance high-frequency information has been proven, and extensively adopted in low-level (e.g., ISR) and high-level computer vision tasks [35,36,37,38].
Motivated by ISR in the visible spectral domain, Choi et al. [39] proposed TEN, a network consisting of four convolutional layers for end-to-end mapping of LR thermal images. Limited by the expensive HR thermal detectors at that time, it was difficult to obtain a large number of paired LR-HR thermal images, TEN used RGB images as the training dataset. Similarly, Marivani et al. [40] proposed multimodal SR models DMSC and DMSC + , using HR RGB images as an auxiliary to super-resolve LR near-infrared thermal images. However, Rivadeneira et al. [41] demonstrated that a network trained with thermal images is superior for thermal image SR inference compared to RGB training datasets. In addition, the authors also constructed a dataset consisting of 101,640 × 512 resolution thermal images to activate the research in the field of thermal image enhancement. Aiming at the problem of severe thermal image noise caused by high clutter in the maritime environment, Bhattacharya et al. [42] proposed two CNN-based networks to perform denoising and SR tasks, respectively, to improve the perception of maritime thermal images. The idea of cascading two networks is also adopted in CDN_MRF [43]. The first residual network of CDN_MRF is used to extract thermal image structure, and the second network is used for fine high-frequency details. As the champion of the Perception Beyond the Visible Spectrum (PBVS)-2020 Thermal Image SR Challenge, the progressive feature extraction module of the TherISuRNet [2] was adopted to generate HR thermal images. In order to reduce redundant features extraction in the deep networks, ChaSNet [44] with the channel separation method eliminates overload features in the trunk of the network. However, the limited receptive fields of the convolution kernels adopted in the above models limit the performance. Zhang et al. [1] proposed MPRANet, a thermal image network composed of residual blocks with parallel convolution kernels of different sizes to effectively extract local and global features.
Apart from discriminative models, some research efforts have focused on generative models, such as generative adversarial networks (GANs) [45] for thermal image SR. Liu et al. [46] integrated the gradient prior knowledge of natural scenes, and trained the GAN-based thermal image SR network with RGB images as style feature auxiliary information. Rivadeneira et al. [47] proposed CycleGAN, a network based on [48] and with an unsupervised training method. In general, GAN-based thermal image SR model training is unstable and prone to mode collapse, so most thermal image SR methods are still based on CNN.

2.2. Attention Mechanisms in Image Super-Resolution

ISR networks extensively employ CA and SA modules to enhance channel and spatial dimension feature maps, respectively. These two attention paradigms focus on rich patterns to reconstruct HR images.
Channel Attention. Channel attention is divided into scalar-based CA [27] and covariance-based CA [49]. The schematic diagram of the two paradigms is shown in Figure 1. Scalar-based CA generates weights in the channel dimension for reweighting feature maps. This means that all pixels of each channel are rescaled by the same scalar, i.e., CA is the channel dimension isotropy operator. Whereas covariance-based CA performs inner product (i.e., self-attention) operation on the input features to generate a cross-covariance matrix to transfer information between channels.
Spatial Attention. Spatial attention can be viewed as an anisotropic operator, i.e., each pixel of all channels is multiplied by various weights to highlight features. Spatial gate-based [29,30] and self-attention-based [50,51] SA are two representative paradigms. As shown in Figure 1, the spatial gate SA generates channel-independent weight masks, while the self-attention SA computes the cross-covariance in the pixel domain. The information interaction of the above two SAs is only in the spatial dimension, and there is no interaction between the channels.
As shown in Figure 1, the CA operation is isotropic in the spatial dimension, that is, the computational complexity of CA is less than that of SA. Therefore, lightweight ISR networks tend to adopt CA rather than SA. However, some research works [1,21] have shown that the combined of CA and SA can help improve ISR performance. The aggregation capabilities of CA and SA in different feature dimensions complement each other. We propose to use pixel-wise attention, which is more effective than SA, and combined with CA, to improve performance while maintaining a compact network structure.

3. Motivation

According to ISR research works on RGB sensors [18], LR images are mixed with low-frequency information and high-frequency details, such as edges and textures. A network without incorporating an attention mechanism handles all frequency bands equally at all layers, resulting in inefficient redundant computation. The attention mechanism can enhance high-frequency features and improve visual quality. However, there are few studies on thermal image SR task for infrared sensors to prove the above assumptions.
We construct a network composed of attention modules to explore the properties of the attention mechanism in thermal image SR task. The LR thermal image first passes through a convolutional layer to extract shallow features, then sequentially extracts deep features by 16 attention blocks, and finally outputs the HR thermal image through an upsampling module. The proposed attention module consists of sequential channel and spatial attention blocks such that each pixel in the feature map can be rescaled independently. We visualize the feature maps of the outputs of some attention modules as shown in Figure 2. As shown in Figure 2, the behaviors of attention modules at different depths are quite different, even diametrically opposite. The shallow attention modules extract low-frequency features, i.e., flat regions are enhanced, the tail attention modules enhance high-frequency information, and the middle modules mix the above two operations.
Furthermore, to verify whether all attention modules contribute to the final performance gain, we replace the attention modules of some layers with residual blocks, and the results are shown in Table 1. The peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) are performance metrics. Table 1 shows that the attention module actually improves the ISR accuracy, and the location of the attention module is critical to the performance. The pure residual block structure achieved the same performance as the structure with attention modules in the first half, and the quantitative comparison results between the network with attention blocks in the second half and the fully attention network were consistent. According to the above experiments, we can conclude the following conclusions: (1) It is beneficial to adopt the attention mechanism for ISR task. (2) It is not always optimal to embrace the attention blocks equally. Therefore, we propose a dynamic attention network for infrared sensors to super-resolve LR thermal images.

4. Proposed Method

We describe the details of the proposed lightweight dynamic attention super-resolution network (LDASRNet) for infrared sensors in this section. The structure diagram of LDASRNet is shown in Figure 3.
The LDASRNet consists of the following three modules:
(1)
Shallow feature extraction (SFE) module. The input LR thermal image first passes through an SFE module consisting of a 3 × 3 convolution kernel to extract low-level features. Our experiments show that using a single 3 × 3 convolution is a acceptable balance between performance and parameters.
(2)
Deep feature extraction (DFE) module. The output of the SFE module is used to extract deeper features through the DFE module. As a key component of LDASRNet, the DFE module consists of K dynamic attention blocks for dynamically enhancing high-frequency feature extraction according to the input feature maps.
(3)
Feature reconstruction (FRec) module. The FRec module constructs the outputs of the DFE module into the final HR thermal image. Due to the lower LR spatial resolution of the ×4 scale factor compared to the ×2 and ×3 scale factors, reconstruction is more difficult. We designed two types of FRec modules specifically for the ×2/×3 and ×4 scale factors, respectively.

4.1. Shallow Feature Extraction Module

Given a LR thermal image I L R R C × H × W , where C is the number of channels, H and W are the height and width, respectively. The function of the SFE module can be expressed mathematically as:
x 0 = f S F E ( I L R )
where x 0 is the output of the SFE module and f S F E ( ) represents the SFE function, which consists of a single 3 × 3 convolution kernel. An alternative scheme is to utilize more convolutional layers to form the SFE module, but we found that this is inefficient and unnecessary. A single 3 × 3 convolutional layer can already extract low-frequency information well, and it balances ISR accuracy and computational burden.

4.2. Deep Feature Extraction Module

As shown in Figure 3, the outputs x 0 of the SFE module are then used to extract high-level patterns through a DFE module composed of K DABs. Each DAB consists of a dynamic weight block (DWB) as well as an attention branch and a non-attention branch. The structure of DAB is shown in Figure 4.

4.2.1. Dynamic Weights Block

As mentioned in Section 3, simply stacking attention blocks failed lead to optimal performance gains. Therefore, we propose a dynamic attention mechanism to maximize the enhancement of high-frequency information. The specific implementation of the dynamic attention mechanism is shown in Figure 4.
The input x i 1 R C i 1 × H i 1 × W i 1 passes through the global average pooling (GAP) layer of i-th DAB to obtain a feature vector z i 1 R C i 1 × 1 × 1 , where the k-th statistic of z i 1 is calculated according to:
z i 1 ( k ) = f a v g ( x i 1 ( k ) ) = 1 H i 1 × W i 1 m = 1 H i 1 n = 1 W i 1 x i 1 ( k ) ( m , n )
where f a v g ( ) is channel-wise GAP, x i 1 ( k ) R 1 × H i 1 × W i 1 and z i 1 ( k ) R 1 × 1 × 1 are the feature map of the k-th channel and its corresponding GAP output, respectively. There are many other sophisticated methods for aggregating global information; we use the simplest GAP to achieve this goal efficiently. Intuitively, the usual approach is to follow z i 1 with two fully connected layers, namely a channel reduction layer and a channel increase layer, to enhance the information interaction between channels [27]. However, we use 1D convolution to simplify the above operation and achieve superior performance while reducing the number of parameters. We will demonstrate the necessity of choosing 1D convolution instead of two fully connected layers.
If we choose two fully connected layers, the output w can be expressed as:
w = σ ( f { w 1 , w 2 } ( g ( z i 1 ) ) )
where w R C i 1 × 1 × 1 , σ is a sigmoid function, and the specific form of f { w 1 , w 2 } is:
f { w 1 , w 2 } ( y ) = W 2 RELU ( W 1 y )
where RELU is the rectified linear unit [52]. As the weights of the channel reduction layer W 1 and channel increase W 2 have sizes C × C r and C r × C respectively, r is an adjustable attenuation parameter. The above operation reduces the parameter burden but destroys the direct one-to-one correspondence between channels and weights [53]. One weight element of a fully connected layer utilizes all channel information; however, the operation shown in Equation (4) first maps full-size features to a low-dimensional space, and then maps back to a high-dimensional space. The direct relationship between channels and weights is broken. We show that 1D convolution elegantly preserves the explicit correspondence between channels and weights.
We enhance the cross-channel interaction using 1D convolution as shown in the following equation:
w = σ softmax ( f FC ( f C 1 D t ( z i 1 ) ) )
where f C 1 D t ( ) represents a 1D convolution with filter size t, and f FC ( ) and σ softmax ( ) are fully connected layer and softmax functions, respectively.
Our goal is to generate two weights for the attention and non-attention branches respectively from the input feature maps. We use GAP based on the following two considerations: (1) As the depth of the network increases, GAP effectively increases the receptive field, thereby extracting global image information. (2) Compared with directly applying the fully connected layer, the GAP drastically reduces parameters, suppresses overfitting and can flexibly adapt to changes in the size of input feature maps. Therefore, the outputs of our proposed DWB block are:
w attn , w n attn = σ softmax ( f FC ( RELU ( f C 1 D t ( z i 1 ) ) ) )
We use w attn and w n attn to recalibrate the attention branch and no-attention branch, respectively. The above operation can be formulated as:
x i = f 1 × 1 ( w i 1 attn × x i 1 attn + w i 1 n ttn × x i 1 n attn ) + x i 1
where w i 1 attn and w i 1 n ttn represent the weights of the attention branch and the non-attention branch, respectively. x i 1 attn is the output of the attention branch, and x i 1 n attn is the output of the non-attention branch. f 1 × 1 ( ) is a convolution layer with a convolution kernel size of 1 × 1.
In order to reduce the learning difficulty and compress the filter space, we let w attn + w n attn = 1 , and the softmax function obtains the normalized weights.
The size of the 1D convolution kernel t determines the interaction range between local channels. In order to avoid manually determining t, which consumes time and resources, we automatically determine t in the following way:
t = ϕ ( C i 1 ) = | log 2 ( C i 1 ) γ + b γ | o d d
where | z | o d d represents the odd number closest to z, C i 1 denotes the number of channels, γ and b are two hyper-parameters, and we empirically set γ = 2 and b = 1 , respectively. We use nonlinear mapping instead of linear mapping to extend the representation capacity, and the local interaction range is proportional to the channel dimension size.

4.2.2. Attention Branch

As shown in Figure 4, the attention branch consists of a channel attention block (CAB) and pixel attention block (PAB), and their structures are shown in Figure 5 and Figure 6b, respectively.
Similar to DWB, we use GAP to obtain channel-by-channel global information, and then perform cross-channel interaction with 1D convolution without channel dimensionality reduction. Our efficient CAB reduces model complexity while capturing local dependencies between channels.
The characteristics of feature map of each channel vary greatly. As shown in Figure 6a, the conventional spatial attention block equally weights all pixels of each channel feature map, which cannot fully enhance the spatial information. Inspired by PAN [54], we propose pixel attention with residual connections, which weights the pixels of each channel independently as shown in Figure 6b.

4.2.3. Non-Attention Branch

As a complement, we introduce the non-attention branch to extract information ignored by the attention branch. We use a single 3 × 3 convolutional layer to form the non-attention branch. It is worth noting that non-attention branches with more complex structures can be adopted as alternatives, but 3 × 3 convolutional layers are suitable for our proposed lightweight structure.
To sum up, the output of the DFE module is:
x n = f D A B K 1 ( f D A B K 2 ( f D A B 1 ( f D A B 0 ( x 0 ) ) ) )
where f D A B i ( ) , i = 0 , 1 , , K 1 , is the i-th DAB.

4.3. Feature Reconstruction Module

The FRec module is used to reconstruct the outputs of the DFE module into the final HR thermal image. There are few works that carefully design upsampling modules. We design two FRec modules for ×2/×3 and ×4 scale factors, respectively, as shown in Figure 7.
We use nearest neighbor interpolation in the FRec module to upsample the feature maps to the desired size, and leverage PAB to enhance information representation. Since the ISR task with ×4 scale factor is more burdensome, we designed the FRec module as shown in Figure 7b for the ×4 scale factor.
Overall, the generated HR thermal image output I S R can be expressed as:
I S R = f F R e c ( x n ) + f u p ( I L R )
where f F R e c ( ) and f u p ( ) represent the FRec module and bilinear interpolation, respectively. We interpolate the LR thermal image to the desired size, allowing the network to learn the residual information, thereby reducing the burden and stability of the network training.

5. Experimental Analysis

5.1. Training and Testing Datasets

We use the dataset proposed by Rivadeneira et al. [47] as the training dataset. This dataset was serviced as the training and testing dataset for the PBVS [55] Thermal Image SR (TISR) challenge, which we simply abbreviate as the Challenge dataset. The Challenge dataset was created by capturing thermal images from three thermal cameras mounted on a panel. The panel was installed on the car and controlled by a developed multi-threaded script to acquire images simultaneously. The specific specifications of the three thermal cameras and the composition of the Challenge dataset are shown in Table 2 and Table 3, respectively.
Since the medium-resolution (MR) Axis and LR Domo thermal images of the Challenge dataset are not completely aligned, we adopt the Flir HR subdataset as the training dataset, and the corresponding LR counterparts are obtained by the bicubic interpolation method. As for the testing datasets, in addition to the Challenge testing dataset, in order to reflect the superior generalization of the proposed LDASRNet, we also handle Iray (http://iray.iraytek.com:7813/apply/E_Super_resolution.html/, accessed on 20 September 2023) and FLIR (https://www.flir.in/oem/adas/adas-dataset-form/, accessed on 20 September 2023) as two additional testing datasets.

5.2. Evaluation Metrics

PSNR and SSIM are adopted to quantitatively evaluate the performance of the proposed LDASRNet and compared models. All performance reports are evaluated on the Y channel of YCbCr color space. Following previous research [1], we crop s pixels around the generated I S R , where s = 2 , 3 , 4 is the corresponding scale factor.

5.3. Data Augmentation Method

Previous studies have shown that feature domain data augmentation (DA) methods harm the performance of ISR task, while pixel domain DA methods boost ISR accuracy. Inspired by CutBlur [56] and MPRANet [1], we adopt a mixture of DA (MoDA) strategy when training the proposed LDASRNet. Specifically, in addition to random horizontal/vertical flipping and rotation of the LR-HR pairs in each iteration during the training, one of the following pixel domain DA methods is also randomly selected to enhance the LR-HR image pairs: CutMixup [56], RGB permute [56], Blend [56], CutBlur [56], CutOut [57], CutMix [58] and Mixup [59]. Quantitative performance comparison of the proposed LDASRNet with various DA methods is shown in Table 4.
Table 4 demonstrates that the pixel domain DA methods can effectively improve ISR performance. For example, compared with the baseline model, the proposed LDASRNet trained using RGB permute or Mixup improved the PSNR metric by at least 0.09 dB, while using the CutMixup or CutBlur method improved it by 0.12 dB. Furthermore, we obtained the highest performance gains using the MoDA strategy. The results show that the pixel domain MoDA strategy in the ISR task can effectively boost accuracy.

5.4. Implementation Details

We train the proposed LDASRNet using the PyTorch [19] framework. We adopt the AdamW [60] optimizer instead of Adam [61] optimizer, and ablation experiments show that AdamW can slightly improve performance compared to Adam. There are 2000 epochs in total, and the initial learning rate 5 × 10 4 decreases by half every 200 epochs. The cropped ground truth resolutions corresponding to ×2, ×3 and ×4 scale factors are 96 × 96, 128 × 128 and 192 × 192, respectively.
We propose a variant network of LDASRNet named LDASRNet-T. The only difference from LDASRNet is that the non-attention branch in LDASRNet-T consists of a convolution layer with a convolution kernel size 1 × 1.

5.5. Ablation Experiments

AdamW vs. Adam. We find that LDASRNet trained with the AdamW optimizer achieves higher scores on PSNR and SSIM metrics compared to Adam. The results are shown in Table 5. We speculate that this is because AdamW directly uses the weight decay term when updating the weights.
Table 5 shows that LDASRNet trained with the AdamW optimizer on the Challenge, FLIR and Iray testing datasets achieved performance improvements of 0.01 dB, 0.04 dB and 0.08 dB, respectively, in the PSNR metric compared to Adam. Moreover, the MoDA training strategy achieves improved accuracy with AdamW or Adam optimizer, further confirming the rationality of MoDA for ISR task.
Validity of DAB structure. To verify the effectiveness of the proposed DAB structure, our ablation experiments compare the impact of a single path and two paths (i.e., attention branch and non-attention branch) on ISR performance. All experiments were performed on the Challenge testing dataset with the ×4 scale factor, and the results are shown in Table 6. Table 6 shows that, only equipped with non-attention branch or attention branch, the PSNR metric is 0.28 dB and 0.51 dB lower than LDASRNet, respectively. This justifies attention and non-attention branches to complement each other. Note that using only the non-attention branch results in a 0.23 dB improvement compared to using the attention branch, but the number of parameters of the former is 2.7 times that of the latter, which shows that our attention branch composed of CAB and PAB is an efficient lightweight structure.
In addition to the addition we used, the fusion methods of the two branches also include strategies such as concatenation [54] and adaptive weight [18]. We compare the performance and parameter trade-offs of the adopted addition versus concatenation and adaptive weight. As shown in case 3 and case 5 in Table 6, concatenation has 25.6 K more parameters than the additive and adaptive weight fusion methods but has the worst performance. We adopt the simplest addition strategy to achieve the best accuracy while maintaining the smallest scope of parameters.
Model capacity. The capacity of the model, i.e., the width and depth of the network, is critical to ISR accuracy. The number of filters in each convolutional layer in the DAB of LDASRNet is 40. In order to ablate the impact of the model capacity on performance, we propose two variant networks, LDASRNet w/Fewer Channels and LDASRNet-T. There are 32 feature channels in the DAB of LDASRNet w/Fewer Channels. The number of channels of LDASRNet-T remains the same as LDASRNet, but the kernel size of the convolutional layer in the non-attention branch is 1 × 1. Case 6 and case 8 in Table 6 show that fewer channels or a small convolution kernel in the non-attention branch is not conducive to the final thermal image SR performance. Our LDASRNet achieved optimal performance while remaining lightweight.
Configuration of two-path structure. We show the ablation experimental results in Table 7 to verify the impact of different configurations of the two-path structure. The two pairs of configurations, case 1 and case 5, and case 4 and case 7, show that non-attention branch equipped with CAB or PAB with dynamic attention block failed always bring positive gains. And the case 6 results indicate that dynamically assigned weights to CAB and PAB are 0.2 dB higher than the cascaded CAB and PAB structure, and the dynamic attention strategy can obtain clear performance improvements. Our non-attention branch and attention branch are supplemented with the dynamic weight structure to achieve superior accuracy.

5.6. Quantitative Experiments

The proposed LDASRNet is a lightweight ISR network, and we choose the following networks with parameters less than 1M for comparison: SRCNN [15], FSRCNN [16], SR-LUT [62], PAN [54], DRRN [63], A 2 F [64], AWSRN-S [65], IMDN [66], VDSR [32] and A 2 N [18]. To verify that LDASRNet has comparable or even better performance than larger networks, we select models AWSRN [65], SRMDNF [67], CARN [68], ChaSNet [44], MPRANet [1] and MDSR [26] with parameters ranging from 1.4M to 6.5M for comparison. In addition, we select RCAN [24] and EDSR [26], two networks with parameters exceeding 10M (EDSR has more than 40M parameters) as reference. The qualitative measurement results of LDASRNet and the above models on the Challenge, FLIR and Iray datasets are shown in Table 8. We show the comparison results of the three metrics PSNR, SSIM and FLOPs.
The quantitative comparison results in Table 8 show that the proposed LDASRNet achieves the highest PSNR and SSIM and the lowest complexity, i.e., the smallest FLOPs, on the three testing datasets with ×2, ×3 and ×4 scale factors. On the Challenge dataset with ×3 scale factor, for instance, LDASRNet is 0.12 dB higher than the second-best A 2 N , while the scope of parameters is only about one-third of it (0.34 M vs. 1.0 M).
Surprisingly, we still achieve superior performance compared to networks with a size scope larger than LDASRNet. Specifically, the networks identified with † in Table 8 represent parameters larger than LDASRNet but the performance is worse. ChaSNet with ×4 scale factor has 14.5 M parameters, which is 41 times that of LDASRNet, and the PSNR metric is 2.16 dB lower than ours. Even compared to EDSR with more than 40M parameters, LDASRNet is only, at most, 0.2 dB lower than it on the Challenge dataset.
We also adopt an edge preservation index (EPI) metric to measure the degree of high-frequency detail recovery in the super-resolved HR thermal images. A higher EPI score indicates a higher perceived quality of the generated HR thermal image. The EPI measurement results of LDASRNet and compared models are shown in Table 9. As shown in Table 9, our LDASRNet obtains the highest EPI metric, meaning that the proposed LDASRNet can recover the most complete high-frequency details.

5.7. Qualitative Experiments

The qualitative comparison results are shown in Figure 8, Figure 9 and Figure 10. Due to space limitations, we perform qualitative experiments on the Iray, FLIR and Challenge testing datasets with ×2, ×3 and ×4 scale factors, respectively. Note that since ChaSNet only provides model structures with ×2 and ×4 scale factors, the comparison results of ×3 exclude ChaSNet. As can be seen from Figure 8, Figure 9 and Figure 10, the proposed LDASRNet not only obtains the best PSNR and SSIM metrics, but also recovers the most complete details compared to the compared networks.
Specifically, Figure 8 shows on the thermal image img0005 from the Iray test dataset with the ×2 scale factor that our LDASRNet achieves a PSNR metric that is 0.11 dB higher than the second-best model AWSRN, i.e., 36.33 dB vs. 36.22 dB. Similarly, Figure 9 exhibits that LDASRNet outperforms the second-best and third-best networks by 0.05 dB/0.0012 and 0.09 dB/0.0027 in terms of the PSNR/SSIM metrics, respectively, and still achieves the highest quantitative measurements. Figure 10 indicates that LDASRNet achieves the maximum performance gain on the ×4 scale factor of the Challenge test dataset for the img0037, with a PSNR of 35.22 dB and SSIM of 0.9195, outperforming A 2 N and AWSRN by 0.31 dB and 0.0042, respectively.

5.8. Compare with Lucy–Richardson–Rosen Algorithm

In addition, we further compare the proposed LDASRNet with the recent Lucy–Richardson–Rosen algorithm (LRRA) [69,70] that exhibits superior deblurring performance. Empirically, we set the maximum number of iterations of LRRA to 8, and we use a synthetic Gaussian-type point-spread function (PSF) with standard deviation 0.3 and filter of size 3 × 3. The quantitative test results of our LDASRNet and LRRA on the ×2, ×3 and ×4 scale factors of the three test datasets of Challenge, FLIR and Iray are shown in Table 10. The comparison results of the visual perception quality of the generated images are shown in Figure 11. Note that due to space limitations, our results only show img0011 from Iray, img0006 from FLIR and img0035 from Challenge.
Table 10 shows that our LDASRNet achieves the highest metrics compared to LRRA on the three test datasets with all scale factors. Specifically, on Challenge with ×2, ×3 and ×4 scale factors, our LDASRNet is 7.07 dB, 6.4 dB and 6.23 dB higher than LRRA in the PSNR metric, respectively. Similarly, on the FLIR and Iray test datasets, LDASRNet is at least 3.84 dB/0.0782 and 0.98 dB/0.4313 higher than LRRA in PSNR/SSIM metrics, respectively, i.e., 35.48 dB/0.8683 vs. 31.64 dB/0.7901 and 27.35 dB/0.8873 vs. 26.37 dB/0.8563. The results show that the proposed LDASRNet has a higher signal-to-noise ratio and more complete structural recovery than the reconstruction results of LRRA.
Qualitatively, Figure 11 shows that the HR thermal images generated by our LDASRNet possess more high-frequency details than LRRA. For example, for the buildings in img0011 and img0006, the edges and textures in the results of our method are clearer, while the LRRA images are blurry. Similarly, in img0035 in Challenge, the fence generated by LRRA obviously lacks high-frequency details and has poor perceptual quality compared to our result.
Notably, LRRA demonstrated its excellent deblurring performance in a previous study [70]. We believe that the input LR thermal image is almost free of blur, which is the main reason why LRRA is not as expected. Accordingly, comparison with deep learning-based networks and LRRA demonstrate the superior performance of the proposed LDASRNet for thermal image SR. The superior HR thermal image reconstruction accuracy and compact model size of the proposed LDASRNet show the potential for deployment in edge devices.

6. Conclusions

In this paper, we show that simply stacking attention modules at different depths of the deep network is suboptimal. Based on this observation, we propose a lightweight thermal image super-resolution network LDASRNet based on the dynamic attention mechanism for infrared sensors. The dynamic weight block in the proposed LDASRNet provides masks to the attention and non-attention branches according to the input features to enhance high-frequency detail extraction. Specifically, we use 1D convolution without dimensionality reduction to replace the fully connected layer to enrich the interactions between channels. The attention branch consisting of efficient channel attention and pixel attention blocks complements the non-attention branch to extract local and global features. Qualitative and quantitative experiments on three testing datasets containing various scenarios show that the proposed LDASRNet can recover high-frequency details accurately, and the lightweight structure has the potential to be deployed on edge devices.

Author Contributions

Conceptualization, Y.H. and H.Z.; methodology, H.Z.; software, H.Z.; validation, H.Z. and Y.H.; formal analysis, H.Z.; investigation, H.Z.; resources, H.Z.; data curation, H.Z.; writing—original draft preparation, H.Z.; writing—review and editing, H.Z.; visualization, M.Y.; supervision, Y.H.; project administration, Y.H.; funding acquisition, Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 62073210.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, H.; Hu, Y.; Yan, M.; Ma, B. Thermal image super-resolution via multi-path residual attention network. Signal Image Video Process. 2023, 17, 2073–2081. [Google Scholar] [CrossRef]
  2. Chudasama, V.; Patel, H.; Prajapati, K.; Upla, K.P.; Ramachandra, R.; Raja, K.; Busch, C. Therisurnet—A computationally efficient thermal image super-resolution network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 86–87. [Google Scholar]
  3. Wang, Z.; Chen, J.; Hoi, S.C. Deep learning for image super-resolution: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3365–3387. [Google Scholar] [CrossRef] [PubMed]
  4. Park, S.C.; Park, M.K.; Kang, M.G. Super-resolution image reconstruction: A technical overview. IEEE Signal Process. Mag. 2003, 20, 21–36. [Google Scholar] [CrossRef]
  5. Li, X.; Orchard, M.T. New edge-directed interpolation. IEEE Trans. Image Process. 2001, 10, 1521–1527. [Google Scholar]
  6. Zhang, L.; Wu, X. An edge-guided image interpolation algorithm via directional filtering and data fusion. IEEE Trans. Image Process. 2006, 15, 2226–2238. [Google Scholar] [CrossRef] [PubMed]
  7. Mishiba, K.; Suzuki, T.; Ikehara, M. Edge-adaptive image interpolation using constrained least squares. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010; pp. 2837–2840. [Google Scholar]
  8. Dong, W.; Zhang, L.; Shi, G.; Wu, X. Nonlocal back-projection for adaptive image enlargement. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 349–352. [Google Scholar]
  9. Hardie, R.C.; Barnard, K.J.; Armstrong, E.E. Joint MAP registration and high-resolution image estimation using a sequence of undersampled images. IEEE Trans. Image Process. 1997, 6, 1621–1633. [Google Scholar] [CrossRef] [PubMed]
  10. Hardie, R.C.; Barnard, K.J.; Bognar, J.G.; Armstrong, E.E.; Watson, E.A. High-resolution image reconstruction from a sequence of rotated and translated frames and its application to an infrared imaging system. Opt. Eng. 1998, 37, 247–260. [Google Scholar]
  11. Bishara, W.; Su, T.W.; Coskun, A.F.; Ozcan, A. Lensfree on-chip microscopy over a wide field-of-view using pixel super-resolution. Opt. Express 2010, 18, 11181–11191. [Google Scholar] [CrossRef] [PubMed]
  12. Bishara, W.; Sikora, U.; Mudanyali, O.; Su, T.W.; Yaglidere, O.; Luckhart, S.; Ozcan, A. Holographic pixel super-resolution in portable lensless on-chip microscopy using a fiber-optic array. Lab Chip 2011, 11, 1276–1279. [Google Scholar] [CrossRef] [PubMed]
  13. Greenbaum, A.; Ozcan, A. Maskless imaging of dense samples using pixel super-resolution based multi-height lensfree on-chip microscopy. Opt. Express 2012, 20, 3129–3143. [Google Scholar] [CrossRef]
  14. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image super-resolution via sparse representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef] [PubMed]
  15. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef]
  16. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 391–407. [Google Scholar]
  17. Zou, L.; Xu, S.; Zhu, W.; Huang, X.; Lei, Z.; He, K. Improved Generative Adversarial Network for Super-Resolution Reconstruction of Coal Photomicrographs. Sensors 2023, 23, 7296. [Google Scholar] [CrossRef]
  18. Chen, H.; Gu, J.; Zhang, Z. Attention in attention network for image super-resolution. arXiv 2021, arXiv:2104.09497. [Google Scholar]
  19. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic Differentiation in Pytorch. Available online: https://openreview.net/forum?id=BJJsrmfCZ (accessed on 22 October 2023).
  20. Park, J.; Woo, S.; Lee, J.Y.; Kweon, I.S. Bam: Bottleneck attention module. arXiv 2018, arXiv:1807.06514. [Google Scholar]
  21. Wang, H.; Chen, X.; Ni, B.; Liu, Y.; Liu, J. Omni Aggregation Networks for Lightweight Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22378–22387. [Google Scholar]
  22. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  23. Hu, Y.; Li, J.; Huang, Y.; Gao, X. Channel-wise and spatial feature modulation network for single image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 3911–3927. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  27. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  28. Zhang, Y.; Li, K.; Li, K.; Zhong, B.; Fu, Y. Residual non-local attention networks for image restoration. arXiv 2019, arXiv:1903.10082. [Google Scholar]
  29. Chen, L.; Chu, X.; Zhang, X.; Sun, J. Simple baselines for image restoration. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 17–33. [Google Scholar]
  30. Yang, J.; Li, C.; Dai, X.; Gao, J. Focal modulation networks. Adv. Neural Inf. Process. Syst. 2022, 35, 4203–4217. [Google Scholar]
  31. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 84–90. [Google Scholar] [CrossRef]
  32. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  33. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  34. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  35. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 191–207. [Google Scholar]
  36. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.T.; Zhang, L. Second-order attention network for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11065–11074. [Google Scholar]
  37. Liu, D.; Wen, B.; Fan, Y.; Loy, C.C.; Huang, T.S. Non-local recurrent network for image restoration. Adv. Neural Inf. Process. Syst. 2018, 31, 1680–1689. [Google Scholar]
  38. Mei, Y.; Fan, Y.; Zhou, Y. Image super-resolution with non-local sparse attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3517–3526. [Google Scholar]
  39. Choi, Y.; Kim, N.; Hwang, S.; Kweon, I.S. Thermal image enhancement using convolutional neural network. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 223–230. [Google Scholar]
  40. Marivani, I.; Tsiligianni, E.; Cornelis, B.; Deligiannis, N. Multimodal image super-resolution via deep unfolding with side information. In Proceedings of the 2019 27th European Signal Processing Conference (EUSIPCO), A Coruna, Spain, 2–6 September 2019; pp. 1–5. [Google Scholar]
  41. Rivadeneira, R.E.; Suárez, P.L.; Sappa, A.D.; Vintimilla, B.X. Thermal image superresolution through deep convolutional neural network. In Proceedings of the Image Analysis and Recognition: 16th International Conference (ICIAR 2019), Waterloo, ON, Canada, 27–29 August 2019; Proceedings, Part II 16; Springer: Berlin/Heidelberg, Germany, 2019; pp. 417–426. [Google Scholar]
  42. Bhattacharya, P.; Riechen, J.; Zölzer, U. Infrared Image Enhancement in Maritime Environment with Convolutional Neural Networks. In Proceedings of the VISIGRAPP (4: VISAPP), Madeira, Portugal, 27–29 January 2018; pp. 37–46. [Google Scholar]
  43. He, Z.; Tang, S.; Yang, J.; Cao, Y.; Yang, M.Y.; Cao, Y. Cascaded deep networks with multiple receptive fields for infrared image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 2310–2322. [Google Scholar] [CrossRef]
  44. Prajapati, K.; Chudasama, V.; Patel, H.; Sarvaiya, A.; Upla, K.P.; Raja, K.; Ramachandra, R.; Busch, C. Channel split convolutional neural network (ChaSNet) for thermal image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4368–4377. [Google Scholar]
  45. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  46. Liu, S.; Yang, Y.; Li, Q.; Feng, H.; Xu, Z.; Chen, Y.; Liu, L. Infrared image super resolution using gan with infrared image prior. In Proceedings of the 2019 IEEE 4th International Conference on Signal and Image Processing (ICSIP), Wuxi, China, 19–21 July 2019; pp. 1004–1009. [Google Scholar]
  47. Rivadeneira, R.E.; Sappa, A.D.; Vintimilla, B.X. Thermal Image Super-resolution: A Novel Architecture and Dataset. In Proceedings of the VISIGRAPP (4: VISAPP), Valletta, Malta, 27–29 February 2020; pp. 111–119. [Google Scholar]
  48. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  49. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5728–5739. [Google Scholar]
  50. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. [Google Scholar]
  51. Lu, Z.; Li, J.; Liu, H.; Huang, C.; Zhang, L.; Zeng, T. Transformer for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 457–466. [Google Scholar]
  52. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  53. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar]
  54. Zhao, H.; Kong, X.; He, J.; Qiao, Y.; Dong, C. Efficient image super-resolution using pixel attention. In Proceedings of the Computer Vision—ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; Proceedings, Part III 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 56–72. [Google Scholar]
  55. Rivadeneira, R.E.; Sappa, A.D.; Vintimilla, B.X.; Bin, D.; Ruodi, L.; Shengye, L.; Zhong, Z.; Liu, X.; Jiang, J.; Wang, C. Thermal Image Super-Resolution Challenge Results-PBVS 2023. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 470–478. [Google Scholar]
  56. Yoo, J.; Ahn, N.; Sohn, K.A. Rethinking data augmentation for image super-resolution: A comprehensive analysis and a new strategy. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8375–8384. [Google Scholar]
  57. DeVries, T.; Taylor, G.W. Improved regularization of convolutional neural networks with cutout. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  58. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
  59. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. Mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  60. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  61. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  62. Jo, Y.; Kim, S.J. Practical Single-Image Super-Resolution Using Look-Up Table. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 691–700. [Google Scholar]
  63. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3147–3155. [Google Scholar]
  64. Wang, X.; Wang, Q.; Zhao, Y.; Yan, J.; Fan, L.; Chen, L. Lightweight single-image super-resolution network with attentive auxiliary feature learning. In Proceedings of the Asian Conference on Computer Vision, Kyoto, Japan, 30 November–4 December 2020. [Google Scholar]
  65. Wang, C.; Li, Z.; Shi, J. Lightweight image super-resolution with adaptive weighted learning network. arXiv 2019, arXiv:1904.02358. [Google Scholar]
  66. Hui, Z.; Gao, X.; Yang, Y.; Wang, X. Lightweight image super-resolution with information multi-distillation network. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
  67. Zhang, K.; Zuo, W.; Zhang, L. Learning a single convolutional super-resolution network for multiple degradations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3262–3271. [Google Scholar]
  68. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
  69. Anand, V.; Han, M.; Maksimovic, J.; Ng, S.H.; Katkus, T.; Klein, A.; Bambery, K.; Tobin, M.J.; Vongsvivut, J.; Juodkazis, S. Single-shot mid-infrared incoherent holography using Lucy-Richardson-Rosen algorithm. Opto-Electron. Sci. 2022, 1, 210006-1. [Google Scholar]
  70. Jayavel, A.; Gopinath, S.; Periyasamy Angamuthu, P.; Arockiaraj, F.G.; Bleahu, A.; Xavier, A.P.I.; Smith, D.; Han, M.; Slobozhan, I.; Ng, S.H.; et al. Improved Classification of Blurred Images with Deep-Learning Networks Using Lucy-Richardson-Rosen Algorithm. Photonics 2023, 10, 396. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of channel attention and spatial attention in different paradigms.
Figure 1. Schematic diagram of channel attention and spatial attention in different paradigms.
Sensors 23 08717 g001
Figure 2. Visualization of feature maps. In the input and output feature maps, white, red, and blue pixels indicate zero, positive, and negative values, respectively. The brighter the pixels in the attention maps, the larger the value of the attention coefficients.
Figure 2. Visualization of feature maps. In the input and output feature maps, white, red, and blue pixels indicate zero, positive, and negative values, respectively. The brighter the pixels in the attention maps, the larger the value of the attention coefficients.
Sensors 23 08717 g002
Figure 3. The structure of the proposed lightweight dynamic attention super-resolution network (LDASRNet).
Figure 3. The structure of the proposed lightweight dynamic attention super-resolution network (LDASRNet).
Sensors 23 08717 g003
Figure 4. The overall structure of the dynamic attention block. AvgPool and FC represent global average pooling and full connection operations, respectively.
Figure 4. The overall structure of the dynamic attention block. AvgPool and FC represent global average pooling and full connection operations, respectively.
Sensors 23 08717 g004
Figure 5. The overall structure of the channel attention block. GAP represents global average pooling, and k = 5 represents the 1D convolution kernel size.
Figure 5. The overall structure of the channel attention block. GAP represents global average pooling, and k = 5 represents the 1D convolution kernel size.
Sensors 23 08717 g005
Figure 6. The overall structures of the spatial and pixel attention blocks. (a) Spatial attention block. (b) The proposed residual pixel attention block.
Figure 6. The overall structures of the spatial and pixel attention blocks. (a) Spatial attention block. (b) The proposed residual pixel attention block.
Sensors 23 08717 g006
Figure 7. The overall structures of the feature reconstruction blocks. NN Inter is nearest neighbor interpolation operation. (a) Feature reconstruction block for ×2 and ×3 scale factors. (b) Feature reconstruction block for ×4 scale factor.
Figure 7. The overall structures of the feature reconstruction blocks. NN Inter is nearest neighbor interpolation operation. (a) Feature reconstruction block for ×2 and ×3 scale factors. (b) Feature reconstruction block for ×4 scale factor.
Sensors 23 08717 g007
Figure 8. The qualitative results on the Iray testing dataset with the ×2 scale factor.
Figure 8. The qualitative results on the Iray testing dataset with the ×2 scale factor.
Sensors 23 08717 g008
Figure 9. The qualitative results on the FLIR testing dataset with the ×3 scale factor.
Figure 9. The qualitative results on the FLIR testing dataset with the ×3 scale factor.
Sensors 23 08717 g009
Figure 10. The qualitative results on the Challenge testing dataset with the ×4 scale factor.
Figure 10. The qualitative results on the Challenge testing dataset with the ×4 scale factor.
Sensors 23 08717 g010
Figure 11. Comparison of the visual quality of thermal images processed by the proposed LDASRNet and LRRA.
Figure 11. Comparison of the visual quality of thermal images processed by the proposed LDASRNet and LRRA.
Sensors 23 08717 g011
Table 1. Attention modules contribute to thermal image SR performance. Attention module index indicates that the i-th position is an attention module and the rest are residual blocks, and PSNR and SSIM are performance metrics. Red and blue text indicate best and second-best performance, respectively.
Table 1. Attention modules contribute to thermal image SR performance. Attention module index indicates that the i-th position is an attention module and the rest are residual blocks, and PSNR and SSIM are performance metrics. Red and blue text indicate best and second-best performance, respectively.
Attention Module IndexPSNR↑ SSIM↑
None32.950.8078
All33.240.8099
1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 32.950.8078
9 , 10 , 11 , 12 , 13 , 14 , 15 , 16 33.240.8099
2 , 4 , 6 , 8 , 10 , 12 , 14 , 16 33.200.8089
Table 2. Thermal camera specifications for creating the Challenge dataset.
Table 2. Thermal camera specifications for creating the Challenge dataset.
Image ResolutionCamera ModelFOVFocal LengthPixel SizeSpectral RangeOperating Temperature Range
Low (LR)Axis Domo P129035.44 mm12 μ m8–14 μ m−30–55 ° C
Mid (MR)Axis Q2901-E359 mm17 μ m8–14 μ m−40–60 ° C
High (HR)FC-632O FLIR3219 mm17 μ m7–13.5 μ m−50–70 ° C
Table 3. The structure and composition of the Challenge dataset.
Table 3. The structure and composition of the Challenge dataset.
PhaseSubdatasetResolutionAmount
TrainingDomo160 × 120951
Axis320 × 240
Flir640 × 480
TestingDomo160 × 12050
Axis320 × 240
Flir640 × 480
Table 4. Quantitative performance comparison of the proposed LDASRNet using various data augmentation methods. We take LDASRNet trained using horizontal/vertical flipping and random rotation without the MoDA strategy as the baseline. Results are reported on the Challenge testing dataset with ×4 scale factor. Red and green text indicate the best metric and the gain relative to the baseline, respectively.
Table 4. Quantitative performance comparison of the proposed LDASRNet using various data augmentation methods. We take LDASRNet trained using horizontal/vertical flipping and random rotation without the MoDA strategy as the baseline. Results are reported on the Challenge testing dataset with ×4 scale factor. Red and green text indicate the best metric and the gain relative to the baseline, respectively.
Data Augmentation MethodPSNR↑ SSIM↑
LDASRNet w/o MoDA (Baseline)36.52 (+0.00)0.9291 (+0.0000)
LDASRNet w/CutOut36.62 (+0.10)0.9300 (+0.0009)
LDASRNet w/CutMix36.63 (+0.11)0.9303 (+0.0012)
LDASRNet w/Mixup36.61 (+0.09)0.9300 (+0.0009)
LDASRNet w/CutMixup36.64 (+0.12)0.9309 (+0.0018)
LDASRNet w/RGB permute36.61 (+0.09)0.9299 (+0.0008)
LDASRNet w/Blend36.60 (+0.08)0.9298 (+0.0007)
LDASRNet w/CutBlur36.64 (+0.12)0.9310 (+0.0019)
LDASRNet w/MoDA (Our)36.65 (+0.13)0.9312 (+0.0021)
Table 5. Performance comparison between AdamW and Adam optimizers. All results are reported on the ×2 scale factor. Red text indicates the best metrics.
Table 5. Performance comparison between AdamW and Adam optimizers. All results are reported on the ×2 scale factor. Red text indicates the best metrics.
AdamAdamWMoDAChallenge DatasetFLIR DatasetIray Dataset
PSNR↑SSIM↑PSNR↑SSIM↑PSNR↑SSIM↑
44.180.977435.290.858333.260.9358
44.200.978835.330.9376
44.330.978535.440.868933.400.9384
44.340.987035.480.868333.480.9466
Table 6. Ablation studies: effects of different DAB structural configurations. LDASRNet w/Fewer Channels means that the number of channels of the feature map in DAB is reduced to 32. All experiments were performed on the Challenge testing dataset with the ×4 scale factor. Red text indicates the best metrics.
Table 6. Ablation studies: effects of different DAB structural configurations. LDASRNet w/Fewer Channels means that the number of channels of the feature map in DAB is reduced to 32. All experiments were performed on the Challenge testing dataset with the ×4 scale factor. Red text indicates the best metrics.
PathPath FusionParams (K)Metric
Non-AttentionAttentionAdditionConcatenationAdaptive-WeightLDANet (1 × 1 conv)LDANet (3 × 3 conv)PSNR↑
Single PathLDASRNet w/Non-Attention 323.29936.37
LDASRNet w/Attention 119.18736.14
Two Paths w/o DABLDASRNet w/Addition 349.58736.50
LDASRNet w/Concatenation 375.18736.47
LDASRNet w/Adaptive-Weight 349.58736.46
LDASRNetLDASRNet w/Fewer Channels 232.81136.58
LDASRNet-T 146.13136.60
LDASRNet 350.93136.65
Table 7. Ablation studies: effects of various structural configurations of two paths. C-P AB means channel- and pixel-wise attention block. All experiments were performed on the Challenge testing dataset with ×4 scale factor. Red text indicates the best metric.
Table 7. Ablation studies: effects of various structural configurations of two paths. C-P AB means channel- and pixel-wise attention block. All experiments were performed on the Challenge testing dataset with ×4 scale factor. Red text indicates the best metric.
SettingsParams (K)MetricGains from DAB
Non-AttentionCABPABC-P ABDABPSNR↑PSNR↑
323.34736.40
119.18736.17
349.58736.53
349.53936.51
324.69136.38+0.02
120.53136.19−0.02
350.88336.60+0.09
350.93136.65+0.12
Table 8. Quantitative comparison results of PSNR/SSIM/FLOPs metrics. Red and blue texts indicate the best and second-best performance, respectively (except those with parameters greater than 10 M). † indicates that the model parameters are larger than our LDASRNet but the accuracy is worse.
Table 8. Quantitative comparison results of PSNR/SSIM/FLOPs metrics. Red and blue texts indicate the best and second-best performance, respectively (except those with parameters greater than 10 M). † indicates that the model parameters are larger than our LDASRNet but the accuracy is worse.
ScaleSize ScopeNetworkPramsTesting Datasets
ChallengeFLIRIray
× 2 ≤1 MFSRCNN0.012 M42.11/0.9841/0.0010T34.95/0.8637/0.0010T32.20/0.9361/0.0003T
SR-LUT0.017 M41.23/0.9783/0.0039T34.49/0.8563/0.0041T30.96/0.9200/0.0014T
SRCNN0.057 M39.22/0.9790/0.0176T34.79/0.8607/0.0187T31.83/0.9314/0.0063T
PAN0.1 M39.06/0.9649/0.0142T33.25/0.8205/0.0152T28.85/0.8804/0.0051T
DRRN0.3 M43.44/0.9848/0.8161T35.34/0.8663/0.8705T32.92/0.9423/0.2938T
A 2 F 0.3 M44.26/0.9868/0.0236T35.43/0.8676/0.0251T33.25/0.9451/0.0085T
AWSRN-S0.4 M43.89/0.9861/0.0304T35.38/0.8670/0.0324T33.03/0.9438/0.0109T
IMDN0.7 M44.12/0.9861/0.0015T35.46/0.8672/0.0016T33.36/0.9451/0.0006T
VDSR0.7 M43.77/0.9868/0.2042T35.36/0.8679/0.2178T32.89/0.9438/0.0735T
A 2 N 1.0 M44.23/0.9867/0.0826T35.47/0.8679/0.0881T33.28/0.9450/0.0297T
Ours0.34 M44.34/0.9870/0.0295T35.48/0.8683/0.0315T33.48/0.9466/0.0106T
<7 MAWSRN1.4 M44.38/0.9871/0.1068T35.53/0.8680/0.1140T33.60/0.9474/0.0385T
SRMDNF †1.5 M44.14/0.9869/0.1146T35.41/0.8671/0.1222T33.18/0.9445/0.0412T
CARN †1.6 M43.97/0.9867/0.0743T35.43/0.8676/0.0792T33.35/0.9459/0.0267T
ChaSNet †3.2 M32.73/0.9564/0.2433T30.27/0.8184/0.2592T25.93/0.8580/0.0876T
MPRANet4.4 M45.50/0.9903/0.3892T35.88/0.8857/0.4152T33.99/0.9537/0.1401T
MDSR †6.5 M40.66/0.9804/0.4993T34.40/0.8573/0.5325T30.49/0.9168/0.0799T
>10 MRCAN15.4 M44.50/0.9873/1.1766T35.63/0.8697/1.2708T33.91/0.9496/0.4289T
EDSR40.7 M44.44/0.9872/3.1282T35.56/0.8687/3.3368T33.81/0.9490/1.1261T
× 3 ≤1 MFSRCNN0.012 M37.68/0.9487/0.0004T32.15/0.7958/0.0003T28.29/0.8562/0.0002T
SR-LUT0.017 M37.12/0.9410/0.0017T32.14/0.7858/0.0019T27.79/0.8415/0.0006T
SRCNN0.057 M37.68/0.9505/0.0176T32.54/0.7980/0.0186T28.32/0.8584/0.0063T
PAN0.1 M35.00/0.9249/0.0089T31.08/0.7582/0.0094T26.43/0.8038/0.0032T
DRRN0.3 M38.78/0.9586/0.8148T33.07/0.8075/0.8657T29.22/0.8811/0.2938T
A 2 F 0.3 M39.38/0.9617/0.0105T33.15/0.8082/0.0112T29.39/0.8848/0.0038T
AWSRN-S0.5 M38.97/0.9549/0.0162T33.05/0.8067/0.0172T29.19/0.8819/0.0058T
IMDN0.7 M39.38/0.9621/0.0007T33.17/0.8082/0.0008T29.34/0.8858/0.0003T
VDSR0.7 M38.76/0.9594/0.2039T35.03/0.8075/0.2166T29.17/0.8814/0.0735T
A 2 N 1.0 M39.44/0.9620/0.0392T33.17/0.8086/0.0417T29.41/0.8854/0.0141T
Ours0.34 M39.56/0.9628/0.0157T33.24/0.8099/0.0167T29.50/0.8873/0.0057T
<7 MAWSRN1.5 M39.60/0.9630/0.05101T33.24/0.8103/0.0532T29.64/0.8912/0.0181T
SRMDNF †1.5 M39.14/0.9603/0.0514T33.10/0.8075/0.0546T29.33/0.8841/0.0185T
CARN †1.6 M39.31/0.9617/0.0395T33.18/0.8092/0.0420T29.49/0.8878/0.0143T
ChaSNet
MPRANet4.4 M40.72/0.9741/0.2113T33.34/0.8190/0.2245T30.62/0.9092/0.0762T
MDSR6.7 M36.67/0.9942/0.1365T32.00/0.7881/0.1456T27.40/0.8361/0.0463T
>10 MRCAN15.6 M39.87/0.9646/0.5287T33.33/0.8120/0.5617T30.00/0.8966/0.1906T
EDSR43.7 M39.76/0.9640/1.4899T33.30/0.8118/1.5830T24.73/0.6140/0.5372T
× 4 ≤1 MFSRCNN0.012 M35.11/0.9102/0.0002T31.01/0.7439/0.0003T26.58/0.7922/0.0001T
SR-LUT0.017 M34.59/0.8994/0.0010T30.64/0.7308/0.0011T26.10/0.7748/0.0003T
SRCNN0.057 M35.05/0.9107/0.0175T30.96/0.7435/0.0187T26.45/0.7901/0.0063T
PAN0.2 M33.18/0.8813/0.0071T29.79/0.7029/0.0075T24.97/0.7353/0.0025T
DRRN0.3 M35.97/0.9139/0.8161T31.48/0.7574/0.8705T27.14/0.8176/0.2938T
A 2 F 0.3 M36.45/0.9289/0.0060T31.59/0.7603/0.0064T27.36/0.8266/0.0022T
AWSRN-S0.6 M36.21/0.9263/0.0112T31.51/0.7585/0.0120T27.21/0.8210/0.0040T
IMDN0.7 M36.51/0.9290/0.0005T31.62/0.7607/0.0005T27.31/0.8275/0.0002T
VDSR0.7 M35.98/0.9249/0.2042T31.46/0.7579/0.2178T27.15/0.8197/0.0735T
A 2 N 1.0 M36.40/0.9280/0.0242T31.58/0.7592/0.0258T27.23/0.8233/0.0087T
Ours0.35 M36.65/0.9312/0.0109T31.68/0.7625/0.0116T27.35/0.8296/0.0039T
<7 MAWSRN †1.6 M36.37/0.9316/0.0304T31.67/0.7630/0.0324T27.53/0.8343/0.0109T
SRMDNF †1.5 M36.25/0.9270/0.0294T31.53/0.7591/0.0314T27.33/0.8260/0.0106T
CARN †1.6 M36.48/0.9296/0.0303T31.63/0.7616/0.0323T27.40/0.8292/0.0109T
ChaSNet †14.5 M34.49/0.8957/0.2769T30.55/0.7162/0.2954T26.25/0.7722/0.0997T
MPRANet4.41 M36.95/0.9365/0.1494T31.92/0.7700/0.1593T27.82/0.8488/0.0538T
MDSR †6.7 M34.23/0.9032/0.1366T30.51/0.7336/0.1456T27.40/0.8361/0.0492T
>10 MRCAN15.6 M36.94/0.9244/0.2978T31.74/0.7647/0.3177T27.72/0.8407/0.1072T
EDSR43.7 M36.81/0.9331/1.4899T33.30/0.8118/1.5830T27.70/0.7922/0.5372T
Table 9. Edge preservation index (EPI) metric measurement. All experiments were performed on the Challenge testing dataset with ×2 scale factor. Red and blue text represent the best and second-best metrics respectively.
Table 9. Edge preservation index (EPI) metric measurement. All experiments were performed on the Challenge testing dataset with ×2 scale factor. Red and blue text represent the best and second-best metrics respectively.
ModelBicubicFSRCNNSR-LUTSRCNNPAN A 2 F IMDN A 2 N AWSRNSRMDNFChaSNetMDSROurs
EPI↑0.75490.80690.68200.77430.75870.84560.84420.84700.84780.84280.75020.75170.8497
Table 10. Quantitative comparison results of the proposed LDASRNet and LRRA. PSNR (dB) and SSIM are the metrics.
Table 10. Quantitative comparison results of the proposed LDASRNet and LRRA. PSNR (dB) and SSIM are the metrics.
ScaleMethodTesting Datasets
ChallengeFLIRIray
× 2 LRRA37.27/0.967331.64/0.790132.35/0.9119
Ours44.34/0.987035.48/0.868333.48/0.9466
× 3 LRRA33.16/0.908629.36/0.686728.30/0.8563
Ours39.56/0.962833.24/0.809929.50/0.8873
× 4 LRRA30.42/0.845227.80/0.606026.37/0.7883
Ours36.65/0.931231.68/0.762527.35/0.8296
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Hu, Y.; Yan, M. Thermal Image Super-Resolution Based on Lightweight Dynamic Attention Network for Infrared Sensors. Sensors 2023, 23, 8717. https://doi.org/10.3390/s23218717

AMA Style

Zhang H, Hu Y, Yan M. Thermal Image Super-Resolution Based on Lightweight Dynamic Attention Network for Infrared Sensors. Sensors. 2023; 23(21):8717. https://doi.org/10.3390/s23218717

Chicago/Turabian Style

Zhang, Haikun, Yueli Hu, and Ming Yan. 2023. "Thermal Image Super-Resolution Based on Lightweight Dynamic Attention Network for Infrared Sensors" Sensors 23, no. 21: 8717. https://doi.org/10.3390/s23218717

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop