Next Article in Journal
Prediction of Particulate Matter 2.5 Concentration Using a Deep Learning Model with Time-Frequency Domain Information
Previous Article in Journal
Application of 4 × 44 Oligo Microarray to Transcriptomic Analysis of Immune Response in Rainbow Trout Infected with Aeromonas salmonicida
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale CNN-Transformer Dual Network for Hyperspectral Compressive Snapshot Reconstruction

Engineering Research Center of Digital Forensics, Ministry of Education, Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(23), 12795; https://doi.org/10.3390/app132312795
Submission received: 26 October 2023 / Revised: 19 November 2023 / Accepted: 22 November 2023 / Published: 29 November 2023

Abstract

:
Coded aperture snapshot spectral imaging (CASSI) is a new imaging mode that captures the spectral characteristics of materials in real scenes. It encodes three-dimensional spatial–spectral data into two-dimensional snapshot measurements, and then recovers the original hyperspectral image (HSI) through a reconstruction algorithm. Hyperspectral data have multi-scale coupling correlations in both spatial and spectral dimensions. Designing a network architecture that effectively represents this coupling correlation is crucial for enhancing reconstruction quality. Although the convolutional neural network (CNN) can effectively represent local details, it cannot capture long-range correlation well. The Transformer excels at representing long-range correlation within the local window, but there are also issues of over-smoothing and loss of details. In order to cope with these problems, this paper proposes a dual-branch CNN-Transformer complementary module (DualCT). Its CNN branch mainly focuses on learning the spatial details of hyperspectral images, and the Transformer branch captures the global correlation between spectral bands. These two branches are linked through bidirectional interactions to promote the effective fusion of spatial–spectral features of the two branches. By utilizing characteristics of CASSI imaging, the residual mask attention is also designed and encapsulated in the DualCT module to refine the fused features. Furthermore, by using the DualCT module as a basic component, a multi-scale encoding and decoding model is designed to capture multi-scale spatial–spectral features of hyperspectral images and achieve end-to-end reconstruction. Experiments show that the proposed network can effectively improve reconstruction quality, and ablation experiments also verify the effectiveness of our network design.

1. Introduction

Hyperspectral imaging technology can capture the spectral information of real-world scenes within a specific wavelength range. Compared with three-channel visible RGB images, hyperspectral images usually have a wider spectral response range and a finer spectral sampling resolution. Each pixel represents spectral signatures of objects in the scene and can be used to better identify their categories. Therefore, hyperspectral images are widely used in various tasks, including ground object classification [1,2,3,4], target detection and tracking, and change detection [2,5]. In order to obtain complete spatial and spectral information, conventional hyperspectral imaging technology needs to scan in spectral or spatial dimensions. However, the scanning operation requires more imaging time and is not suitable for capturing dynamic scenes. Recently, a new hyperspectral imaging technology, called snapshot compressive imaging (SCI), has emerged; it is based on compressive sensing theory, which can significantly improve imaging efficiency. In these SCI systems, coded aperture snapshot spectral imaging (CASSI) is a potential solution. The CASSI system modulates incoming light through a coded aperture and accumulates the modulated light on a two-dimensional sensor array into a two-dimensional snapshot measurement during a single exposure time. Then the 3D hyperspectral image (HSI) cube can be recovered from the 2D snapshot measurement during the reconstruction stage. The design of the reconstruction algorithm is the core issue of the CASSI system and a key factor affecting imaging quality.
Traditional reconstruction methods [6,7,8,9] recover the original hyperspectral image by solving optimization problems with prior regularization constraints. Some commonly used prior constraints include sparsity [10,11,12], total variation [8,10,13] and non-local similarity [9,14,15], etc. However, these handcrafted regularization constraints cannot adequately represent the spatial–spectral structure in hyperspectral images, which degrades the reconstruction quality. At the same time, the reconstruction procedure requires an iterative solution and the time complexity is high. In recent years, CNNs have been widely used in computer vision-related tasks due to their excellent capability of feature learning. They can adaptively learn the hierarchical feature representation of images in a data-driven way, and the learned features can be regarded as the deep prior of the underlying images. To this end, the researchers introduce CNN into the task of compressive snapshot imaging and guide the network to learn the end-to-end mapping from 2D snapshot measurements to 3D HSI. The CNN-based method has brought an effective improvement in reconstruction quality. However, due to the small receptive field of convolution kernels, CNN has certain limitations in capturing non-local self-similarity and long-range correlations in hyperspectral images. The Transformer is a new network structure that uses multi-head self-attention to model non-local correlations. It has received much attention in the field of computer vision and is used in image classification [16,17,18,19,20], target detection [21,22,23,24], semantic segmentation, etc. Researchers have also proposed a Transformer-based network for compressive snapshot reconstruction. A representative Transformer-based reconstruction model (dubbed as MST) calculates self-similarity in the spectral dimension, effectively utilizing global similarity between spectral bands for reconstruction. MST [25] can achieve better reconstruction performance than the competing CNN. However, it has shortcomings in the representation of small-scale spatial structures, as the reconstructed images have the drawback of being over-smooth. A single CNN or Transformer structure cannot effectively represent the multi-scale spatial–spectral correlation of hyperspectral images. Therefore, how to effectively represent the coupling correlation between the spatial and spectral dimensions of hyperspectral images is a key issue in improving reconstruction quality.
In order to cope with the above-mentioned issues, we propose a dual-branch CNN-Transformer complementary module (DualCT). As shown in Figure 1, the CNN branch of DualCT focuses more on obtaining the spatial dimension information of hyperspectral images, while the Transformer branch of DualCT captures the global correlation between spectral bands. At the same time, we also introduce bidirectional cross-branch interactions to fuse the complementary features from the spectral-domain Transformer and spatial-domain convolution. Therefore, DualCT can effectively represent the spatial–spectral correlation in hyperspectral images. Meanwhile, considering the physical mechanism of the CASSI imaging mode, we introduce residual mask attention into the DualCT module. It uses the aperture mask to guide the network to focus on regions with high-fidelity spectral representation, thereby enhancing the spatial–spectral structure of the reconstructed hyperspectral images. Based on the DualCT module, we further build a multi-scale encoding and decoding reconstruction network (DualCT-Net) to learn the end-to-end reconstruction mapping from snapshot measurements to the original hyperspectral image. We conduct comparative experiments and ablation studies to verify the performance of the proposed network. The experimental results show that the proposed network can effectively improve reconstruction performance, and the ablation studies also verify the effectiveness of our network design. The main contributions of this paper include:
(1) The proposed DualCT module parallelizes the spectral Transformer branch and spatial CNN branch, and promotes complementary fusion of spatial and spectral features through bidirectional cross-branch interactions between the two branches. Therefore, the DualCT module can effectively integrate the advantages of CNN and Transformer to capture the spatial–spectral features of hyperspectral images.
(2) Taking into account the physical mechanism of compressive snapshot imaging, we designed a residual mask attention mechanism to enhance the DualCT module. It can guide the DualCT module to focus on spatial regions with high-fidelity spectral representation, thereby enhancing the structural details of the reconstructed hyperspectral images.
(3) By using DualCT as the basic building block, we further construct a multi-scale encoding and decoding model to learn the inverse reconstruction mapping. The experimental results verify the effectiveness of the proposed reconstruction network.

2. Related Work

The compressive snapshot imaging system uses reconstruction algorithms to recover the original 3D hyperspectral images from the 2D snapshot measurements. According to the different models used, current reconstruction algorithms can mainly be divided into three types, namely prior regularization-based methods, deep convolutional network-based methods, and Transformer-based methods. Additionally, since we need to design mask attention based on the CASSI imaging process, we also introduce the CASSI imaging model.

2.1. Prior Regularized Reconstruction Method

Compressive snapshot reconstruction is an ill-posed inverse problem. The early work formulates it as a prior regularized optimization problem consisting of a prior regularization term and a data fidelity term. GAP-TV [8] uses total variation as a prior regularization constraint, and DeSCI [14] uses sparse representation and non-local self-similarity as a priori representation of hyperspectral images. However, these prior regularizations cannot effectively represent the complex spatial–spectral structure in hyperspectral images. At the same time, these reconstruction models require iterative optimization. Using the alternating direction multiplier method (ADMM) as an example, each iteration needs to solve the proximal operators of the prior regularization, and the computational complexity is high.

2.2. Deep Convolutional Network Reconstruction Method

In view of the excellent feature learning capabilities of deep networks, researchers have started using deep networks to directly learn the reconstruction mapping from snapshot measurements to original hyperspectral images [26,27,28,29,30,31]. Under this framework, Xiong et al. [32] used convolutional neural networks to learn hyperspectral compressed snapshot reconstruction. Miao et al. [30] proposed a two-stage hyperspectral image reconstruction model called λ -Net. The first stage of λ -Net employs a generative adversarial network (GAN) to obtain an initial reconstruction. The second stage is to refine each spectral band of the initial reconstruction. Meng et al. [33] expanded the iterative steps of the GAP optimization algorithm into different stages of the network model, and its denoising sub-steps were implemented using a pre-trained denoiser. TSA-Net [29] introduces a self-attention mechanism to capture the dependence between spatial and spectral dimensions of hyperspectral images. These methods based on deep networks can directly output the reconstruction results through a feed-forward network calculation without iteration. Therefore, they have high reconstruction efficiency. However, although the above-mentioned convolutional networks can effectively represent local details, they still have certain limitations in capturing long-term spatial–spectral correlation.

2.3. Transformer Reconstruction Method

The Transformer is widely used in natural language processing tasks [34]. Through long-range correlation matching and interaction between image feature tokens, the Transformer has also been introduced into computer vision tasks and brought effective performance improvement, including image classification [16,17,19,35], object detection [21,22,23] and segmentation [36,37,38]. The Transformer also shows good performance in low-level visual tasks [4,39,40,41]. Chen et al. [22] utilized a standard Transformer to construct the backbone model [34] for various image restoration tasks. Liu et al. [40] used the Swin Transformer to build a residual network, and achieved state-of-the-art results in image restoration tasks. Cai et al. [25] proposed a Transformer-based HSI reconstruction model named MST. It treats each spectral band as a token and computes self-attention in the spectral dimension, improving the reconstruction quality effectively. However, hyperspectral images exhibit coupled spatial–spectral correlations, and the MST model is insufficient at representing spatial correlations, which hinders the reconstruction of small-scale spatial structures.

2.4. CASSI Imaging Model

Figure 2 shows the schematic diagram of the CASSI system. The hyperspectral image corresponding to the physical scene is denoted as F R H × W × N λ , where H, W, N λ represent the height, width, and the number of spectral bands, respectively. The CASSI system first modulates F by the aperture mask M * , which is defined as follows:
F ( : , : , n λ ) = F ( : , : , n λ ) M * , 1 n λ N λ ,
where F R H × W × N λ denotes the modulated HSI, ⊙ represents the element-wise dot multiplication. After modulation, the disperser applies a shearing transformation to the modulated hyperspectral image. Taking the Y-axis shift as an example, the transformation process can be expressed as follows:
F ( u , v , n λ ) = F ( x , y + d ( λ n λ c ) , n λ ) ,
where λ c represents the reference spectral band, and d ( λ n λ c ) denotes the translation amount of the n λ -th spectral band in the y direction, typically a linear function of the number of spectral index. All the spectral bands after shear transformation are projected upon the two-dimensional sensor array, and the images of different spectral bands at the same position are accumulated, which can be calculated as follows:
Y = n λ = 1 N λ F ( : , : , n λ ) + N ,
where Y R H × ( W + d ( N λ 1 ) ) represents the captured two-dimensional snapshot measurement, and N R H × ( W + d ( N λ 1 ) ) represents the measurement noise of the sensor during the imaging process. Equation (3) represents the forward observation equation of CASSI imaging. Based on this observation equation, the reconstruction algorithm recovers the original hyperspectral image from the two-dimensional snapshot measurements.

3. Materials and Methods

3.1. Multi-Scale CNN-Transformer Complementary Reconstruction Network

In order to better represent the spatial–spectral coupling correlations in hyperspectral images, we integrate the advantages of CNN and the Transformer and design a dual-branch CNN-Transformer complementary module (DualCT). Using DualCT as the basic module, we further construct a reconstruction network called DualCT-Net to learn the reconstruction mapping from snapshot measurements to the original hyperspectral image. Similar to the multi-scale structure in [25], DualCT-Net is configured as an encoder, a bottleneck module, and a decoder. As shown in Figure 1, the encoder in DualCT-Net takes the snapshot measurement Y as input. Through the reverse dispersion process, the snapshot measurement Y is backward-shifted to obtain the initial reconstruction H 0 R H × W × N λ , and H 0 is mapped into features X 0 R H × W × C through a 3 × 3 convolutional process. X 0 then flows into the multi-scale feature extraction stage. Each scale of the encoder contains two DualCT modules and one downsampling operation. The downsampling operation employs a convolutional layer with a stride of 4 × 4 stride to reduce the spatial dimensions while doubling the number of channels. The output features of the third scale in the encoder enter the bottleneck module and then flow into the decoder. The bottleneck contains two DualCT modules.
The decoder of DualCT-Net is designed as a symmetrical structure of the encoder. Each scale of the decoder has two DualCT modules and one upsampling module, and the upsampling is realized as a deconvolution operation. The same spatial resolution scale of the encoder and decoder are connected through skip connections, which convey features from encoder to decoder. The skip connections are beneficial for reducing the information loss caused by downsampling operations. We denote the output of the highest resolution scale of the decoder as X 0 R H × W × C . After a 3 × 3 convolution processing upon X 0 , we can obtain the reconstruction R 0 R H × W × N λ . Due to the presence of global residual connections in DualCT-Net, R 0 represents the incremental relative to the initial reconstruction H 0 . Finally, the reconstructed hyperspectral image H R H × W × N λ can be obtained by adding R 0 and H 0 . The specific structure of the DualCT module is introduced in detail below.

3.2. DualCT Module

CNN and Transformer have their own advantages in feature learning. In order to effectively represent the coupling of spatial and spectral features of HSIs, our DualCT module deploys the self-attention mechanism of the Transformer and localized convolution as two parallel branches, and uses bidirectional interactions to provide the fusion of complementary features in both spectral and spatial dimensions. At the same time, motivated by the mask guidance mechanism proposed in [25], we designed mask attention and added it to the DualCT module. According to the imaging characteristics of the CASSI system, mask attention can guide the network to focus on regions with high-fidelity HSI representation, enhancing the representation capability of spatial–spectral structural details. Specifically, as shown in Figure 1b, the DualCT module mainly consists of layer normalization (LN) [42], a parallel complementary fusion structure, and mask attention. The parallel complementary fusion structure contains a spatial domain CNN branch and a spectral domain Transformer branch. The two branches achieve effective feature fusion through bidirectional cross-linking interactions. The calculation process of DualCT can be expressed as follows:
X ^ l = C T ( L N ( X l ) ) + X l , X ^ l + 1 = C O N V ( L N ( X ^ l ) ) M + X ^ l ,
where X l represents the feature input to the DualCT module, LN stands the layer normalization, and C T ( · ) represents the complementary fusion of dual branches. M R H × W × C is the mask attention mapping, ⊙ represents element-wise multiplication, and X ^ l + 1 represents the output feature of the DualCT module. We denote C O N V ( L N ( X ^ l ) ) as X l + 1 . The specific design of the complementary fusion structure will be introduced in detail below.

3.2.1. Parallel Complementary Fusion

As shown in Figure 3, the parallel complementary fusion structure consists of a spatial domain CNN branch (abbreviated as Spa-CNN), a spectral domain Transformer branch (abbreviated as Spe-Trans), and bidirectional cross-linking between these two branches. In the Spa-CNN branch, it cascades three consecutive convolution operations with kernel sizes of 1 × 1 , 5 × 5 , and 1 × 1 . The GELU layer is used after each convolution. The Spe-Trans branch exploits the Transformer to capture the global correlation between spectral bands, and its detailed design will be described in the next sub-section. Bidirectional cross-linking facilitates feature fusion between two branches, effectively capturing spatial–spectral correlations in the hyperspectral image. Specifically, the cross-linking is composed of spatial interaction and channel interaction. The channel interaction directs feature information from the Spa-CNN branch to the Spe-Trans branch, enhancing the feature modeling capability in the channel dimension. At the same time, the spatial interaction directs features from the Spe-Trans branch to the Spa-CNN branch, prompting the model to capture the spatial dependencies between different locations. The calculation flow of bidirectional interactions can be expressed as follows:
F c = S p a C N N ( L N ( X l ) ) , A = C I ( F c ) , F t = S p e T r a n s ( X l , A ) , F c = S I ( F t ) F c ,
where X l is the input of the parallel complementary fusion structure, F c is the feature output of the Spa-CNN branch, C I ( · ) is the channel interaction function, SI is the spatial interaction function, and A is the channel attention calculated from F c by C I ( · ) . The spectral-Trans branch takes the X l and attention weight A as input and uses A to weight the value matrix. The detailed formulation of spectral-Trans branch will be presented in Section 3.2.2. Using the output F t of the spectral-Trans branch as input, we calculate the spatial attention map and use it to weigh F c . Therefore, this parallel complementary structure can promote the interaction and fusion of spatial and spectral features.
With regard to channel interaction, we first use a global average pooling layer to obtain the global average feature vector of F c . Then, the global average feature is processed through two consecutive 1 × 1 convolutional layers, and the GELU activation function [43] is used for non-linear transformation between the two convolutional layers. Finally, the channel attention map is generated along the channel dimension by applying the sigmoid function. The specific operations for channel interaction are defined as follows:
A = C I ( F c ) = σ ( C o n v ( G L ( C o n v ( G P ( F c ) ) ) ) ) ,
where σ represents the Sigmoid function, C o n v ( · ) represents the 1 × 1 convolution, GL represents the GELU activation function, and GP represents the global average pooling layer. The channel attention, A, is used to weight each spectral band of the value matrix in the spectral Transformer.
With regard to spatial interaction, it calculates the spatial attention map from F t to weight the output F c of the spatial-domain CNN branch. Specifically, it consists of two 1 × 1 convolutional layers, and the GELU activation function [43] is used between these two convolutional layers. Finally, the spatial attention map is generated by applying the sigmoid function. The specific operations for spatial interaction are defined as follows:
S I ( F t ) = σ ( C o n v ( G L ( C o n v ( F t ) ) ) ) ,
where S I ( · ) represents the spatial interaction process, σ represents the Sigmoid function, C o n v ( · ) represents 1 × 1 convolution, and GL represents the GELU activation function.
According to Equation (5), we can obtain the attention-enhanced feature F t and F c . The bidirectional cross-branch interactions can effectively enhance the modeling capabilities of the Transformer and deep convolution in both spectral and spatial dimensions, providing complementary features for the two branches. F t , F c are combined through a concatenation operation and then flow into the continuous feed-forward network. FFN consists of two linear layers with a GELU layer in the middle. The final output feature X ^ l of parallel complementary fusion is calculated as follows:
X ^ l = F F N ( C o n c a t ( F c , F t ) ) + X l .

3.2.2. Spectral Domain Transformer Branch

Due to the effectiveness of the Transformer in capturing long-range dependencies, we utilize the Transformer branch to capture the global correlation information between spectral bands, and name it the spectral Transformer. As in [25], it treats each spectral band as a token and performs self-attention calculations between each token. Figure 4 shows the diagram of the spectral Transformer. The input X l R H × W × C of the spectral Transformer is reshaped into X R H W × C , and each column of X is taken as a token. X is then linearly projected to obtain query Q R H W × C , the key K R H W × C , and the value V R H W × C . They are defined as
Q = X W Q , K = X W K , V = X W V ,
where W Q , W K , W V R C × C are learnable projection matrices. We split Q, K, and V into N heads along the channel dimension, and each head independently calculates self-attention within its channel group which is defined as follows:
S a j = s o f t m a x ( μ j K j T Q j ) , V j = A V j , h e a d j = V j S a j ,
where j is the index of head number, μ j R is the learnable parameter. In order to introduce the complementary feature from the spatial-domain CNN branch, we use the channel interaction to weight the value matrix V with channel attention A. Channel attention A is calculated according to Equation (6). Then the outputs of N heads are concatenated along the spectral dimension to form a larger feature matrix. Considering the spatial location relationship within each token, we also add positional embeddings to integrate positional information into the feature representation. The output of spectral Transformer can be calculated as
S p e T r a n s ( X l ) = ( N j = 1 ( h e a d j ) ) W + f p ( V ) ,
where W R C × C is the linear projection matrix, and ∪ is the concatenation of N heads. The positional embeddings function f p ( · ) consists of two 3 × 3 convolutional layers, a GELU activation layer, and a reshaping operation. Finally, after a reshaping operation, we can obtain the output feature map X o u t R H × W × C .

3.2.3. Residual Mask Attention

In the imaging process of the CASSI system, the original scene is first modulated through an aperture mask, and the aperture mask is usually set to a random matrix with elements 0 and 1. The positions of element 1 in the mask allow the light reflected from the target in the scene to pass through, so these positions have high information fidelity. Based on this imaging mode, the literature [25] designed a mask guidance mechanism to enhance the quality of reconstructed images. Inspired by [25], we designed a residual mask attention mechanism and encapsulated it into the DualCT module. This mask attention can guide the network to focus on spatial regions with high-fidelity spectral representations. As shown in Figure 5, in accordance with the CASSI imaging mechanism, we first perform the reverse dispersion on the aperture mask M * :
M s ( x , y , n λ ) = M * ( x , y + d ( λ n λ c ) ) ,
where M s R H × ( W + d ( N λ 1 ) ) × N λ represents the shifted version of M * . Then M s undergoes a sequence of operations to obtain the two-dimensional weight map. These operations include two consecutive 3 × 3 convolutional layers followed by GELU activation, a 5 × 5 convolutional layer and a sigmoid activation function. With the residual link, we can obtain the spatially enhanced mask M s R H × ( W + d ( N λ 1 ) ) × C , and it is calculated as follows:
M s = M s ( 1 + σ ( φ β 2 M s ) ) ,
where σ ( · ) represents the sigmoid activation function, β ( · ) represents the conv 3 × 3 convolution operation and GELU operation, and φ ( · ) represents the mapping function of the 5 × 5 deep convolution layer. We perform the reverse dispersion process and shift M s backward to obtain the mask attention map M R H × W × C :
M ( x , y , n λ ) = M s ( x , y d ( λ n λ c ) , n λ ) ,
where the spectral bands are indexed as n λ [ 1 , , C ] . The obtained mask attention map M adequately indicates the spatial positions with high-fidelity spectral information. Therefore, we integrate the mask attention M into the DualCT module and enhance feature maps in a residual formulation:
X ^ l + 1 = X l + 1 M + X ^ l ,
where X l + 1 = C O N V ( L N ( X ^ l ) ) represents the deep feature map in the DualCT module, X ^ l is the output feature of the parallel complementary fusion structure, and X ^ l + 1 represents the output feature of the DualCT module. The mask attention is helpful for recovering spatial–spectral structures and improving reconstruction performance. Ablation experiments in the next section verify the effectiveness of mask attention.

3.3. Network Parameters Learning

We train the constructed DualCT-Net in a supervised way. We denote Ω as the training dataset. Ω is composed of N data pairs ( x i , y i ) , where y i denotes the compressive snapshot measurements corresponding to the original hyperspectral images, x i . In order to better guide the parameter learning, we design a composite loss function composed of root mean square error (RMSE) loss L R M S E and spectrum constancy loss [44] L S C L . This composite loss function is defined as follows:
L m i x ( θ ) = L R M S E ( θ , Ω ) + γ 1 L S C L ( θ , Ω ) ,
where θ represents the network parameters, and γ 1 denotes the weight parameters that coordinate the importance of the root mean square error and the spectral constant loss. Specifically, the L R M S E loss calculates the root mean square error (RMSE) between the reconstructed images and the original images; the formula is defined as follows:
L R M S E ( θ , Ω ) = 1 N i = 1 N ( x i F N e t ( y i , θ ) ) 2 ,
where F N e t ( y i , θ ) is the hyperspectral image reconstructed from y i by DualCT-Net. Considering the spatial–spectral correlation of hyperspectral images, the spectrum constancy loss L S C L is introduced and is defined as follows:
L S C L ( θ , Ω ) = 1 N i = 1 N λ Ω i λ x i 2 ,
where λ represents the gradient along the spectrum across different spectral bands. This composite loss function can constrain the reconstructed hyperspectral image to approximate in the original domain and gradient domain, resulting in better reconstruction quality.

4. Results

This section verifies the effectiveness of the proposed DualCT-Net through comparative experiments. The proposed network model is quantitatively and qualitatively compared with multiple state-of-the-art methods.

4.1. Experimental Setting

We use CAVE [5] as the training hyperspectral image set. CAVE has 205 hyperspectral images with a spatial size of 1024 × 1024 and 28 spectral bands. The wavelengths of these 28 spectral bands are within the wavelength range of 450–650 nm, and they are 453.3, 457.6, 462.1, 466.8, 471.6, 476.5, 481.6, 486.9, 492.4, 498.0, 503.9, 509.9, 516.2, 522.7, 529.5, 536.5, 543.8, 551.4, 558.6, 567.5, 575.3, 584.3, 594.4, 604.2, 614.4, 625.1, 636.3, and 648.1 nm, respectively. We randomly crop massive sub-images with a spatial size of 256 × 256 from the original large-sized image, thereby enhancing the training image set. According to the forward model of CASSI, we can obtain the snapshot measurement with a spatial size of 256 × 310 , corresponding to each training sample. By performing a reverse dispersion process and backward shift upon the snapshot measurement, we can obtain the reconstruction H 0 and take it as the network input. As in [25], 10 scenes are selected from KAIST [26] for testing.
Furthermore, this paper initialized weights using the method described in [29]. All models were trained for 300 iterations using the Adam [45] optimizer ( β 1 = 0.9 and β 2 = 0.999). We used Adam as the optimizer to train the proposed DualCT-Net, the training was implemented on the PyTorch platform, and CuDNN was used for acceleration. In the network training process, the number of iterations is set to 300, the batch size is set to 5, and the initial learning rate is set to 0.0004, which is halved every 50 epochs.

4.2. Comparative Experiments

In order to evaluate the reconstruction quality, we introduce two quantitative metrics: peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) [46]. Higher values for both metrics indicate better reconstruction quality. We compare the proposed DualCT-Net with four state-of-the-art deep network-based reconstruction methods: TSA-Net [29], SRN [47], GST [48], and MST [25]. A small variant of the MST model is used in our experiments and we denote it as MST-S. Table 1 shows the PSNR and SSIM values of the five reconstruction methods across ten scenes.
As can be seen from Table 1, our DualCT-Net has the best reconstruction results upon all 10 test scenes. TSA-Net [29], SRN [47], and GST [48] are CNN-based reconstruction methods, and MST-S [25] is a Transformer-based reconstruction method. Our network surpasses comparative methods by a large margin. More specifically, the average PSNR value of the DualCT-Net model is 4.34 dB, 3.55 dB, 2.79 dB, and 1.54 dB higher than the state-of-the-art deep network-based methods TSA-Net, SRN, GST, and MST-S, respectively. These results verify the effectiveness of configuring the CNN + Transformer in the DualCT module. The DualCT-Net can effectively integrate the advantages of CNN and the Transformer, capturing spatial–spectral features of hyperspectral images. Figure 6 and Figure 7 visualize spectral bands reconstructed by five methods and the ground-truth spectral bands. We also zoom in on some areas for comparison.
It can be clearly seen that the reconstructed images of TSA-Net can recover the outline of the object, but produce blurry, distorted, and incomplete feature maps. Reconstructed images of SRN and GST are prone to having artifacts along the object contours. Compared with the MST-S method, the DualCT-Net network can reconstruct more clear details.
In addition, Figure 8 and Figure 9 show reconstructed spectral signatures corresponding to annotated rectangular regions in the RGB images. Correlation coefficients are also calculated to quantitatively evaluate the fidelity of reconstructed spectral signatures. Spectral signatures reconstructed by our method can adequately match the reference signatures and have spectral fidelity.

5. Discussion

In this section, we conduct ablation experiments to verify the impact of the parallel complementary fusion structure and mask attention on the reconstruction performance. Additionally, the computational complexity and parameter count of these methods are also analyzed.

5.1. Ablation Experiments

In this section, we conduct ablation experiments to examine the impact of the complementary fusion structure and mask attention on the reconstruction performance. The main verification method is to remove specific modules from the DualCT-Net network and determine their impact on reconstruction performance. The ablation experiments are performed on an HSI dataset [26]. The model obtained by removing our mask guidance mechanism and parallel complementary fusion structure from DualCT-Net is regarded as the baseline model. The ablation experiment results are shown in Table 2. The reconstruction quality of the baseline model is 33.88 dB, and when we use the parallel complementary fusion structure and masked attention, the reconstruction performance is continuously improved by 0.87 dB and 1.05 dB. These experimental results fully demonstrate the effectiveness of the parallel complementary fusion structure and mask attention. Next, we further analyze the impact of different settings within the parallel complementary fusion structure and mask attention on the reconstruction results.

5.1.1. The Ablation of Complementary Fusion Structure

This section continues to verify the effectiveness of parallel branches and bidirectional cross-linking interactions in the DualCT module. The bidirectional cross-linking attention is composed of the channel iteration attention and spatial interaction attention. The experimental results are shown in Table 3.
According to the PSNR and SSIM values in Table 3, we can see that the design of parallel branches and bidirectional cross-linking interactions are both positive factors that can improve reconstruction performance.

5.1.2. Mask Attention Ablation

In this section, we further verify the impact of mask attention on reconstruction performance. Experimental results are shown in Table 4.
Methods A and C utilize the initial reconstruction H 0 , while methods B and D adopt H 0 M * as input. Methods C and D use the Mask Attention in the DualCT module. Method B achieves limited improvements due to the HSI representation damage and insufficient mask utilization. In both cases of input, mask attention can improve the reconstruction performance. These results demonstrate the effectiveness of our mask attention design.

5.2. Computational Complexity and Parameter Amount Analysis

This section further analyzes the computational complexity and parameter amount of the five reconstruction methods. These five methods are all run on NVIDIA GTX 3090 GPU. Table 5 displays the FLOPS and the number of parameters for each reconstruction method. Due to the introduction of bidirectional interactions between the CNN branch and the Transformer branch in the DualCT module, the FLOPS value of our network becomes higher. On the other hand, the DualCT module integrates the advantages of CNN and the Transformer, bringing effective reconstruction performance improvements compared to only using the Transformer or CNN networks.

6. Conclusions

In this paper, we aim to integrate the advantages of CNN and Transformer for representing multi-scale coupling correlations within hyperspectral images. Specifically, we propose a dual-branch CNN-Transformer complementary module. It combines the Transformer self-attention mechanism with deep convolution through dual branches and bidirectional cross-branch interactions. Therefore, it can learn complementary features in both spectral and spatial dimensions. In addition, according to the compressive snapshot imaging mode, this paper also introduces a mask guidance mechanism to refine the spatial–spectral structure of the reconstructed image. By using the DualCT module as a basic component, this paper further designs a multi-scale encoding and decoding model to learn the end-to-end reconstruction mapping from snapshot measurements to the original hyperspectral images. Quantitative experimental results have verified the effectiveness of our network design. It is hoped that the network design and potential insights proposed in this paper will contribute to future work in emerging SCI research.

Author Contributions

Conceptualization: Y.S.; data curation: K.H. and Q.G.; methodology: K.H.; software: K.H.; supervision: Y.S.; writing—original draft: K.H.; writing—review and editing: Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grants 62276139 and U2001211.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study can be available on request from the corresponding author. The data are not publicly available due to privacy.

Acknowledgments

The authors would like to thank the people who shared data and basic models from their research with the community. Their selflessness significantly drives the research process for the entire community. In addition, we sincerely express our gratitude to Huang Junru for her assistance with the experiments and valuable discussions.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Borengasser, M.; Hungate, W.S.; Watkins, R. Hyperspectral Remote Sensing: Principles and Applications; CRC Press: Boca Raton, FL, USA, 2007. [Google Scholar]
  2. Yuan, Y.; Zheng, X.; Lu, X. Hyperspectral image superresolution by transfer learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 1963–1974. [Google Scholar] [CrossRef]
  3. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  4. Wen, J.; Huang, J.; Chen, X.; Huang, K.; Sun, Y. Transformer-Based Cascading Reconstruction Network for Video Snapshot Compressive Imaging. Appl. Sci. 2023, 13, 5922. [Google Scholar] [CrossRef]
  5. Park, J.I.; Lee, M.H.; Grossberg, M.D.; Nayar, S.K. Multispectral imaging using multiplexed illumination. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio De Janeiro, Brazi, 14–21 October 2007; pp. 1–8. [Google Scholar]
  6. Fu, Y.; Zheng, Y.; Sato, I.; Sato, Y. Exploiting spectral-spatial correlation for coded hyperspectral image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3727–3736. [Google Scholar]
  7. Tan, J.; Ma, Y.; Rueda, H.; Baron, D.; Arce, G.R. Compressive hyperspectral imaging via approximate message passing. IEEE J. Sel. Top. Signal Process. 2015, 10, 389–401. [Google Scholar] [CrossRef]
  8. Yuan, X. Generalized alternating projection based total variation minimization for compressive sensing. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 2539–2543. [Google Scholar]
  9. Zhang, S.; Wang, L.; Fu, Y.; Zhong, X.; Huang, H. Computational hyperspectral imaging based on dimension-discriminative low-rank tensor recovery. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10183–10192. [Google Scholar]
  10. Kittle, D.; Choi, K.; Wagadarikar, A.; Brady, D.J. Multiframe image estimation for coded aperture snapshot spectral imagers. Appl. Opt. 2010, 49, 6824–6833. [Google Scholar] [CrossRef]
  11. Lin, X.; Liu, Y.; Wu, J.; Dai, Q. Spatial-spectral encoded compressive hyperspectral imaging. ACM Trans. Graph. 2014, 33, 1–11. [Google Scholar] [CrossRef]
  12. Wagadarikar, A.; John, R.; Willett, R.; Brady, D. Single disperser design for coded aperture snapshot spectral imaging. Appl. Opt. 2008, 47, B44–B51. [Google Scholar] [CrossRef] [PubMed]
  13. Wang, L.; Xiong, Z.; Gao, D.; Shi, G.; Wu, F. Dual-camera design for coded aperture snapshot spectral imaging. Appl. Opt. 2015, 54, 848–858. [Google Scholar] [CrossRef] [PubMed]
  14. Liu, Y.; Yuan, X.; Suo, J.; Brady, D.J.; Dai, Q. Rank minimization for snapshot compressive imaging. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2990–3006. [Google Scholar] [CrossRef]
  15. Wang, L.; Xiong, Z.; Shi, G.; Wu, F.; Zeng, W. Adaptive nonlocal sparse representation for dual-camera compressive hyperspectral imaging. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 2104–2111. [Google Scholar] [CrossRef]
  16. Arnab, A.; Dehghani, M.; Heigold, G.; Sun, C.; Lučić, M.; Schmid, C. Vivit: A video vision transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 6836–6846. [Google Scholar]
  17. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  18. Ali, A.; Touvron, H.; Caron, M.; Bojanowski, P.; Douze, M.; Joulin, A.; Laptev, I.; Neverova, N.; Synnaeve, G.; Verbeek, J.; et al. Xcit: Cross-covariance image transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 20014–20027. [Google Scholar]
  19. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  20. Bazi, Y.; Bashmal, L.; Rahhal, M.M.A.; Dayil, R.A.; Ajlan, N.A. Vision transformers for remote sensing image classification. Remote Sens. 2021, 13, 516. [Google Scholar] [CrossRef]
  21. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 213–229. [Google Scholar]
  22. Dai, X.; Chen, Y.; Yang, J.; Zhang, P.; Yuan, L.; Zhang, L. Dynamic detr: End-to-end object detection with dynamic attention. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2988–2997. [Google Scholar]
  23. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  24. Jiao, J.; Gong, Z.; Zhong, P. Dual-Branch Fourier-Mixing Transformer Network for Hyperspectral Target Detection. Remote Sens. 2023, 15, 4675. [Google Scholar] [CrossRef]
  25. Cai, Y.; Lin, J.; Hu, X.; Wang, H.; Yuan, X.; Zhang, Y.; Timofte, R.; Van Gool, L. Mask-guided spectral-wise transformer for efficient hyperspectral image reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17502–17511. [Google Scholar]
  26. Choi, I.; Kim, M.; Gutierrez, D.; Jeon, D.; Nam, G. High-Quality Hyperspectral Reconstruction Using a Spectral Prior; Technical Report; Association for Computing Machinery: New York, NY, USA, 2017. [Google Scholar]
  27. Hu, X.; Cai, Y.; Lin, J.; Wang, H.; Yuan, X.; Zhang, Y.; Timofte, R.; Van Gool, L. Hdnet: High-resolution dual-domain learning for spectral compressive imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17542–17551. [Google Scholar]
  28. Huang, T.; Dong, W.; Yuan, X.; Wu, J.; Shi, G. Deep gaussian scale mixture prior for spectral compressive imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16216–16225. [Google Scholar]
  29. Meng, Z.; Ma, J.; Yuan, X. End-to-end low cost compressive spectral imaging with spatial-spectral self-attention. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 187–204. [Google Scholar]
  30. Miao, X.; Yuan, X.; Pu, Y.; Athitsos, V. l-net: Reconstruct hyperspectral images from a snapshot measurement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4059–4069. [Google Scholar]
  31. Wang, L.; Sun, C.; Fu, Y.; Kim, M.H.; Huang, H. Hyperspectral image reconstruction using a deep spatial-spectral prior. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8032–8041. [Google Scholar]
  32. Xiong, Z.; Shi, Z.; Li, H.; Wang, L.; Liu, D.; Wu, F. Hscnn: Cnn-based hyperspectral image recovery from spectrally undersampled projections. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 518–525. [Google Scholar]
  33. Meng, Z.; Jalali, S.; Yuan, X. Gap-net for snapshot compressive imaging. arXiv 2020, arXiv:2012.08364. [Google Scholar]
  34. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  35. Dsouza, K.J.; Ansari, Z.A. Histopathology image classification using hybrid parallel structured DEEP-CNN models. Appl. Comput. Sci. 2022, 18, 20–36. [Google Scholar] [CrossRef]
  36. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-like pure transformer for medical image segmentation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 205–218. [Google Scholar]
  37. Wu, B.; Xu, C.; Dai, X.; Wan, A.; Zhang, P.; Yan, Z.; Tomizuka, M.; Gonzalez, J.; Keutzer, K.; Vajda, P. Visual transformers: Token-based image representation and processing for computer vision. arXiv 2020, arXiv:2006.03677. [Google Scholar]
  38. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.; et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 6881–6890. [Google Scholar]
  39. Liu, Z.; Ning, J.; Cao, Y.; Wei, Y.; Zhang, Z.; Lin, S.; Hu, H. Video swin transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3202–3211. [Google Scholar]
  40. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 1833–1844. [Google Scholar]
  41. Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A general u-shaped transformer for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17683–17693. [Google Scholar]
  42. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar]
  43. Hendrycks, D.; Gimpel, K. Gaussian error linear units (gelus). arXiv 2016, arXiv:1606.08415. [Google Scholar]
  44. Zhao, Y.; Guo, H.; Ma, Z.; Cao, X.; Yue, T.; Hu, X. Hyperspectral imaging with random printed mask. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10149–10157. [Google Scholar]
  45. Kinga, D.; Adam, J.B. A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015; Volume 5, p. 6. [Google Scholar]
  46. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  47. Wang, J.; Zhang, Y.; Yuan, X.; Fu, Y.; Tao, Z. A simple and efficient reconstruction backbone for snapshot compressive imaging. arXiv 2021, arXiv:2108.07739. [Google Scholar]
  48. Wang, J.; Zhang, Y.; Yuan, X.; Meng, Z.; Tao, Z. Modeling mask uncertainty in hyperspectral image reconstruction. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 112–129. [Google Scholar]
Figure 1. The architecture of the proposed reconstruction network DualCT-Net. DualCT block acts as the basic building block of DualCT-Net.
Figure 1. The architecture of the proposed reconstruction network DualCT-Net. DualCT block acts as the basic building block of DualCT-Net.
Applsci 13 12795 g001
Figure 2. Schematic diagram of the CASSI system principle.
Figure 2. Schematic diagram of the CASSI system principle.
Applsci 13 12795 g002
Figure 3. Parallel complementary fusion structure inside DualCT module.
Figure 3. Parallel complementary fusion structure inside DualCT module.
Applsci 13 12795 g003
Figure 4. Spectral Transformer.
Figure 4. Spectral Transformer.
Applsci 13 12795 g004
Figure 5. Residual mask attention.
Figure 5. Residual mask attention.
Applsci 13 12795 g005
Figure 6. Visualization of three spectral bands reconstructed by five methods in scene 2.
Figure 6. Visualization of three spectral bands reconstructed by five methods in scene 2.
Applsci 13 12795 g006
Figure 7. Visualization of three spectral bands reconstructed by five methods in scene 8.
Figure 7. Visualization of three spectral bands reconstructed by five methods in scene 8.
Applsci 13 12795 g007
Figure 8. Reconstructed spectral signatures by five methods of the reconstructed hyperspectral image of scene 2.
Figure 8. Reconstructed spectral signatures by five methods of the reconstructed hyperspectral image of scene 2.
Applsci 13 12795 g008
Figure 9. Reconstructed spectral signatures by five methods of the reconstructed hyperspectral image of scene 8.
Figure 9. Reconstructed spectral signatures by five methods of the reconstructed hyperspectral image of scene 8.
Applsci 13 12795 g009
Table 1. PSNR and SSIM values of reconstruction results of 5 methods across 10 scenes.
Table 1. PSNR and SSIM values of reconstruction results of 5 methods across 10 scenes.
   TSA-NetSRNGSTMST-SOurs
Scene 132.03/0.89233.26/0.91033.99/0.92634.71/0.93036.01/0.948
Scene 231.00/0.85829.86/0.88130.49/0.90034.45/0.92537.21/0.953
Scene 332.25/0.91531.69/0.90932.63/0.92135.32/0.94337.66/0.959
Scene 439.19/0.95339.90/0.94741.04/0.96741.50/0.96743.71/0.985
Scene 529.39/0.88430.86/0.92331.49/0.93831.90/0.93333.21/0.958
Scene 631.44/0.90834.20/0.94134.89/0.95533.85/0.94334.95/0.959
Scene 730.32/0.87827.27/0.85227.63/0.86632.69/0.91134.23/0.943
Scene 829.35/0.88832.35/0.93233.02/0.94731.69/0.93332.94/0.954
Scene 930.01/0.89032.83/0.92133.45/0.93234.67/0.93935.70/0.952
Scene 1029.59/0.87430.25/0.90531.49/0.93531.82/0.92632.40/0.942
Average31.46/0.89432.25/0.91233.01/0.92934.26/0.93535.80/0.955
Table 2. Ablation experiments of DualCT-Net. (✓indicates using this module).
Table 2. Ablation experiments of DualCT-Net. (✓indicates using this module).
BaselineParallel Complementary FusionMask AttentionPSNRSSIM
      33.880.931
   34.750.943
35.800.955
Table 3. Ablation experiments of parallel branches and bidirectional cross-linking interactions in the DualCT module. (✓indicates using this module).
Table 3. Ablation experiments of parallel branches and bidirectional cross-linking interactions in the DualCT module. (✓indicates using this module).
Parallel BranchesInteractionPSNRSSIM
ChannelSpatial
         33.880.931
      34.380.939
   34.610.941
   34.610.941
34.750.943
Table 4. Ablation experiments of mask attention in DualCT module. (✓indicates using this module).
Table 4. Ablation experiments of mask attention in DualCT module. (✓indicates using this module).
MethodInputMask AttentionPSNRSSIM
A H 0    34.750.943
B H 0 M *    35.080.946
C H 0 35.800.955
D H 0 M * 35.620.951
Table 5. Analysis of computational complexity and parameter quantity.
Table 5. Analysis of computational complexity and parameter quantity.
AlgorithmParameter Amount (M)FLOPS (G)
TSA-Net44.25110.06
SRN1.2581.84
GST1.2782.87
MST-S0.9312.96
Ours0.92122.27
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, K.; Sun, Y.; Gu, Q. Multi-Scale CNN-Transformer Dual Network for Hyperspectral Compressive Snapshot Reconstruction. Appl. Sci. 2023, 13, 12795. https://doi.org/10.3390/app132312795

AMA Style

Huang K, Sun Y, Gu Q. Multi-Scale CNN-Transformer Dual Network for Hyperspectral Compressive Snapshot Reconstruction. Applied Sciences. 2023; 13(23):12795. https://doi.org/10.3390/app132312795

Chicago/Turabian Style

Huang, Kaixuan, Yubao Sun, and Quan Gu. 2023. "Multi-Scale CNN-Transformer Dual Network for Hyperspectral Compressive Snapshot Reconstruction" Applied Sciences 13, no. 23: 12795. https://doi.org/10.3390/app132312795

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop