Next Article in Journal
Decision Support System Proposal for Medical Evacuations in Military Operations
Next Article in Special Issue
Transformer-Based Semantic Segmentation for Extraction of Building Footprints from Very-High-Resolution Images
Previous Article in Journal
Development and Validation of a Low-Cost Device for Real-Time Detection of Fatigue Damage of Structures Subjected to Vibrations
Previous Article in Special Issue
Generating Defective Epoxy Drop Images for Die Attachment in Integrated Circuit Manufacturing via Enhanced Loss Function CycleGAN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SALSA-Net: Explainable Deep Unrolling Networks for Compressed Sensing

1
School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang 212013, China
2
Jiangsu Engineering Research Center of Big Data Ubiquitous Perception and Intelligent Agriculture Applications, Zhenjiang 212013, China
3
Electronic and Electrical Engineering Department, Brunel University London, Uxbridge UB8 3PH, UK
4
School of Cyberspace Security, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(11), 5142; https://doi.org/10.3390/s23115142
Submission received: 7 May 2023 / Revised: 24 May 2023 / Accepted: 24 May 2023 / Published: 28 May 2023
(This article belongs to the Special Issue Deep Learning-Based Neural Networks for Sensing and Imaging)

Abstract

:
Deep unrolling networks (DUNs) have emerged as a promising approach for solving compressed sensing (CS) problems due to their superior explainability, speed, and performance compared to classical deep network models. However, the CS performance in terms of efficiency and accuracy remains a principal challenge for approaching further improvements. In this paper, we propose a novel deep unrolling model, SALSA-Net, to solve the image CS problem. The network architecture of SALSA-Net is inspired by unrolling and truncating the split augmented Lagrangian shrinkage algorithm (SALSA) which is used to solve sparsity-induced CS reconstruction problems. SALSA-Net inherits the interpretability of the SALSA algorithm while incorporating the learning ability and fast reconstruction speed of deep neural networks. By converting the SALSA algorithm into a deep network structure, SALSA-Net consists of a gradient update module, a threshold denoising module, and an auxiliary update module. All parameters, including the shrinkage thresholds and gradient steps, are optimized through end-to-end learning and are subject to forward constraints to ensure faster convergence. Furthermore, we introduce learned sampling to replace traditional sampling methods so that the sampling matrix can better preserve the feature information of the original signal and improve sampling efficiency. Experimental results demonstrate that SALSA-Net achieves significant reconstruction performance compared to state-of-the-art methods while inheriting the advantages of explainable recovery and high speed from the DUNs paradigm.

1. Introduction

Compressed sensing (CS) [1,2] theory exhibits a novel signal acquisition strategy where a signal can be recovered with overwhelming probability from far fewer acquired measurements than when resolved by the Nyquist sampling theory. This novel signal acquisition paradigm is much more hardware friendly and empowers image capturing with a sub-Nyquist sampling rate [3]. The core idea of CS is to sample and compress at the same time when the recovered signals have sparse representation in some transform domains. Applying CS theory to signal processing can reduce sampling time and cost while ensuring high signal reconstruction accuracy, which is of great significance for many practical applications such as remote sensing [4], single-pixel imaging [5], magnetic resonance imaging (MRI) [6], wireless sensor networks [7], radar imaging [8], spectral compressed imaging [9], computer vision, and pattern recognition [10].
The process of obtaining linear measurements in the theory of CS can be expressed as
y = A x ,
where y R m is the compressed linear measurement, x R n is the original k-sparse (i.e., only k nonzero components in x ) signal to be recovered, and A R m × n is the sampling matrix. The CS sampling rate is defined as r = m / n . In this paper, we adopt a typical block-based image CS problem [11] which splits the high-dimensional natural image into non-overlapped B × B blocks and obtains linear measurements block by block with a relative smaller fixed sampling matrix for the subsequent reconstruction. Since m n , this inverse problem belongs to an underdetermined linear equation. When the sensing matrix A satisfies the constraint of isometry principle [12], the CS reconstruction problem is equivalent to
min x x 0 , s . t . y = A x .
The l 0 -norm minimization problem is typically relaxed to the corresponding l 1 -norm minimization problem, which is a convex optimization problem [13]:
min x x 1 , s . t . y = A x .
By appropriately choosing the regularization parameter λ > 0 , this problem can be further reformulated as an unconstrained optimization problem:
x ^ = arg min x 1 2 y A x 2 2 + λ x 1 .
The compressed sensing theory is dedicated to two main subproblems: signal acquisition and signal reconstruction. The signal acquisition problem focuses on designing efficient sampling matrices to reduce sampling complexity and sampling rate while ensuring reconstruction accuracy. The signal reconstruction problem is concerned with recovering the original signal from the undersampled measurements while satisfying fidelity and stability requirements. We divide existing CS signal reconstruction methods into traditional iterative solutions and deep neural network methods. In this section, we will briefly introduce both of them but focus on the network-based methods most relevant to our own.

1.1. Traditional Methods

In the past decades, various methods [13,14,15] have been developed to solve the problem (4), such as greedy-like algorithms [16], approximate message passing (AMP) [17], proximal gradient descent (PGD) [18], and alternating direction method of multipliers (ADMM) [19]. Greedy-like algorithms iteratively reconstruct a sequence of sparse signals based on support detection and signal recovering using truncated least squares optimization, such as Compressive Sampling Matching Pursuit (CoSaMP) [20], Subspace Pursuit (SP) [21], etc. The most representative PGD-based algorithms include the Iterative Soft Thresholding Algorithm (ISTA) [22], Fast Iterative Soft Thresholding Algorithm (FISTA) [23], and Non-Convex Evolutionary Sparse Target Algorithm (NESTA) [24]. ISTA is a simple iterative thresholding algorithm that uses prior information, such as the sparsity of a signal, to iteratively optimize and reconstruct the signal. FISTA improves upon ISTA by adding an acceleration term, resulting in better reconstruction quality in the same number of iterations. NESTA is a non-convex CS reconstruction algorithm that uses non-convex prior information and has shown good performance in high-dimensional sparse signal reconstruction. ADMM is an optimization method that solves sparse optimization problems by decomposing them into smaller subproblems each of which are then easier to solve. As an instance of ADMM, split augmented Lagrangian shrinkage algorithm (SALSA) [25,26] recasts the CS reconstruction problem into small-scale subproblems that are solved using soft thresholding operators. By iteratively solving the subproblems, SALSA converges significantly faster than ADMM. With well-studied signal formations, these approaches often take the advantage of strong convergence and theoretical analysis. However, they usually give rise to high computational complexity and suffer from choosing optimal prior and tuning parameters.

1.2. Network-Based Methods

Driven by the powerful learning ability of deep neural networks, deep-network-based compressed sensing reconstruction methods have been widely investigated. The core idea of these methods is to learn inverse mapping from compressed measurements simply to map a low-dimensional measurement vector to a high-dimensional image. These methods are usually divided into two categories: one is to train the reconstruction network as a black box classical deep model, and the other is an interpretable deep unrolling model [27] (algorithm unrolling [28] or model-based L2O [29]). In the first category of algorithms, Mousavi et al. [30] proposed a stacked denoising autoencoder (SDA) to solve statistical dependencies between signal elements. However, the fully connected network (FCN) used in SDA produces a large number of learnable parameters. Kulkarni et al. [31] proposed ReconNet, a deep learning model based on convolutional neural networks that captures local image information by directly learning the mapping relationship between compressed sensing measurements and image blocks and then assembling the reconstruction results of each block to achieve image reconstruction. In addition, ReconNet adopts the BM3D [32] algorithm as a denoiser to eliminate block artifacts in the output image. Yao et al. [33] proposed a method called DR 2 -Net, which improves ReconNet by adding residual learning to reconstructed images, achieving better results than ReconNet. Shi et al. [34] proposed a new algorithm called CSNet which learned the compressed sensing reconstruction process through convolutional neural networks while avoiding the problem of manually designing sampling matrices in traditional algorithms, reducing computation and improving reconstruction quality. However, this algorithm has a high-level requirement for training data and is weak in robustness to interference in the case of noisy data. Cui et al. [35] presented NLR-CSNet which aimed to learn a network that can reconstruct images from measurement vectors without pre-training and can achieve good results in low- and high-noise situations. Its non-local adaptive dictionary learning algorithm can learn more representative dictionaries, thus improving the robustness of the network. However, compared with other deep learning models, NLR-CSNet’s model is larger and requires a longer training time. Chen et al. [36] developed CASNet, which uses the adaptive sparse coding (ASC) method to obtain the sparse representation of input data and further improve the effect of sparse representation by using an adaptive threshold mechanism. Zhou et al. [37] proposed BCS-Net, which uses multiple channels to encode different frequency information of the image to improve reconstruction quality. You et al. [38] proposed COAST, which uses a multi-layer convolutional neural network (CNN) to reconstruct measurement values, including an encoder and a decoder. The encoder maps measurement values to a low-dimensional latent space, and the decoder reconstructs the information in the latent space into a complete image. However, the reconstruction speed of the COAST network is relatively slow and is not suitable for real-time applications. These deep models have the advantage of automatically learning features, avoiding the tedious process of manual feature design, and achieving a certain reconstruction effect. However, because they are black box models, they often cannot provide clear predictive explanations, making their use and debugging more difficult. At the same time, they are prone to overfitting when the dataset is small or the model is too complex, leading to training becoming more challenging.
The second type of deep unrolling model combines deep networks with iterative optimization algorithms and exhibits good interpretability. Gregor and LeCun [39] proposed the LISTA deep unrolling model in 2010, which adopts a multi-layer network structure to unroll the ISTA algorithm and uses neural networks to replace the threshold function in the ISTA algorithm for compressed sensing reconstruction of images. Following this seminal work, there has been a surge of efforts [40,41,42,43,44,45] that strive to propose deep unrolling networks by unfolding optimization-based algorithms. These deep unrolling models achieve explainable recovery and high accuracy, which have attracted increasing attention and have become the mainstream for image CS problems. However, these deep learning schemes adopt completely physics-free manners to directly unroll the optimization-based algorithm to learn recovery mapping from the measurements without explicitly making use of sampling processing and physical knowledge. To address these issues, physics-inspired methods incorporate both physical knowledge and sampling processing for further exploration. Zhang et al. [46] proposed the ISTA-Net model with trainable network modules to replace classic ISTA optimization and optimize all network modules through end-to-end learning. This introduced the independent learnable sparsifying/inverse transform with two convolutional layers separated by a Rectified Linear Unit (ReLU). Building on ISTA-Net, You et al. [47] proposed ISTA-Net++, which introduces feature enhancement modules to capture signal features and adds skip connection modules to accelerate network convergence. FISTA-Net [48] directly replaces the general nonorthogonal or even non-linear transform with four convolution layers separated by a ReLU, but no reasonable explanation is given. Yang et al. [49] presented ADMM-CSNet for CS-complex-valued MR imaging problems. The idea behind ADMM-CSNet is to replace the variable splitting and alternating optimization part in the ADMM algorithm with a deep neural network to fully utilize the non-linear mapping ability of neural networks. Liu et al. [50] proposed the RARE model, which uses unsupervised learning to obtain deep prior knowledge for image reconstruction. AMP-Net [51] trains the network using estimation errors generated during the iterative denoising process of the AMP algorithm, resulting in strong generalization performance. MAC-Net [52] introduces memory units in the network and uses double-threshold non-linear mapping and adaptive batch normalization to improve image quality and sparsity. Recently, more flexible backbone networks have emerged. CSformer [53] and TransCS [54] integrate Transformer self-attention-based hybrid architectures to obtain high-quality image recovery.
Interpretable deep unrolling models provide a way to solve image CS problems by incorporating physical knowledge into the model and making the training process more transparent. By absorbing the merits of both physics-free and physics-inspired image CS deep unrolling networks, we propose a novel optimization-based explainable deep unrolling network, coined SALSA-Net. The core idea of the SALSA-Net network is to truncate and unfold the iterations of the SALSA optimization algorithm and map or transform all the steps of each iteration into the end-to-end learning stage; then, all the stages will be concatenated to obtain a unified network. All the parameters involved in SALSA-Net, such as sparsifying/inverse transform, shrinkage threshold, and gradient steps are learned end-to-end. As a result, SALSA-Net takes the advantage of faster convergence and accurate recovery with well-defined explainability.
In summary, the main contributions of this paper are three-fold: (1) A novel deep unrolling model dubbed SALSA-Net is proposed for faster convergence of sparse reconstruction of image CS by mapping the updated steps of SALSA to deep networks. (2) Different from the traditional SALSA algorithm that requires manual tuning of gradient step size and regularization parameters, SALSA-Net learns all the parameters and applies physics-inspired constraints to ensure faster convergence. Furthermore, the sparsifying/inverse transformation in the residual domain is adopted to further improve image reconstruction accuracy. (3) Experimental results show that the proposed method achieves favorable performance against the state-of-the-art approaches in terms of both quantitative measure and visual quality.

2. Method

2.1. SALSA

This section examines the traditional reconstruction algorithm SALSA, which serves as the foundation for SALSA-Net. SALSA is an iterative algorithm designed to solve optimization problems involving both smooth and non-smooth convex functions. The algorithm’s fundamental concept is to divide the objective function into two components: a smooth portion and a non-smooth portion. During the iterative process, the smooth and non-smooth portions are updated separately using gradient descent or conjugate gradient methods, while the Lagrangian multiplier is updated using the augmented Lagrangian multiplier method. More specifically, to solve the problem (4), SALSA recasts it as the following two subproblems:
x ^ = arg min x y A x 2 2 + μ x v + m 2 2 ,
v ^ = arg min v μ 2 x v + m 2 2 + β x 1 ,
where v is a vector with the same dimension as x , m is the Augmented Lagrange multiplier, and μ is a non-negative parameter used to control the weight of the Lagrange term. β is a non-negative parameter used to control the constraint of the L 1 norm. By alternately solving the above two subproblems and updating the Lagrange multiplier m , the reconstruction result of the sparse signal can be iteratively obtained. Specifically, each iteration can be performed according to the following steps:
1.
Augmented Lagrangian term minimization:
v ^ = Soft ( x + m ; β / μ ) ,
where Soft is a threshold function used to shrink the value of x to a nonzero value or zero. Its definition is
Soft ( x , t ) = sign ( x ) max ( | x | t , 0 ) .
2.
Data term minimization:
x ^ = ( μ I + A T A ) 1 ( A T y + μ ( v m ) ) ,
where I is the identity matrix.
3.
Updated the Lagrange multiplier:
m ^ = m v + x ,
where x are the results obtained by data term minimization.
Since SALSA decomposes problems into subproblems that handle only a part of the data at each iteration, it effectively handles large-scale problems. Moreover, SALSA converges faster than other classical algorithms such as ISTA and FISTA [25,26]. The SALSA algorithm has been widely applied in various fields, including image processing, computer vision, and signal processing.

2.2. SALSA-Net

The traditional optimization model SALSA has high computational complexity, and current network models require improvement in restoring image details. To address these issues, this paper proposes a CS image reconstruction network based on a segmentation augmented Lagrangian algorithm. The network framework, shown in Figure 1, is divided into three parts: sampling, initialization, and deep reconstruction. The core algorithm used in this paper differs from ADMM-Net. Additionally, ADMM-Net uses random Gaussian matrix sampling, while SALSA-Net uses convolutional sampling. The structure of the reconstruction module also differs. ADMM-Net is divided into an encoder, decoder, and alternating direction multiplier network, while the reconstruction part of SALSA-Net is divided into a gradient update module, threshold denoising module, and auxiliary update module. We will provide a detailed description of the proposed model in the following subsections.

2.2.1. Sampling

In block-based CS, the image is divided into non-overlapping blocks of size B × B × l , where B × B represents the spatial size of the image block and l represents the number of channels in the image. Each image block is considered an independent signal source, and a sampling matrix A B of size N B × l B 2 is utilized to measure the signals, where N B = r B 2 , r represents the sampling rate. Specifically, the sampling design of CSNet is followed, where each row of the sampling matrix A is treated as a convolution kernel, and a convolution layer is employed to simulate the block-based CS process. This convolution-based sampling method efficiently acquires CS measurements, avoiding the complexity and limitations of traditional hand-designed sampling matrices. The measurement results of each image block are represented as a feature map. This process can be represented using the convolution operator M ( · )
y = M ( x ) .
The input image x is convolved with a non-overlapping convolution operation using a B × B × l -sized convolution kernel with a stride of B × B , which results in the output image y . In accordance with the block-based CS reconstruction method, B is set to 32. During training, the sampling network adaptively learns the sampling matrix, effectively utilizing the local structural features of the image to improve the accuracy and robustness of CS image reconstruction.

2.2.2. Initial Reconstruction

Block-based CS methods use the pseudo-inverse matrix of the sampling matrix to obtain the initial reconstruction of the image, denoted as x = A B ( y ) where A B is the size 1 × 1 × N B . In this paper, a convolution layer and a recombination concatenation layer are used to achieve the initial reconstruction, defined as follows:
x 0 = M ˜ ( y ) ,
where the CS measurements y serve as the input, which undergoes a convolution operation and pixel shuffle operation M ˜ ( · ) to obtain the initial reconstructed image x 0 . Since the output of the sampling network is a 1 × 1 × N B vector, the convolutional kernel size of the initial reconstruction layer is set to 1 × 1 × N B , with a stride of 1 × 1 , to reconstruct each image block independently. Each image block is represented by a vector obtained through a convolutional layer and then recombined through concatenation to form the initial reconstruction image. The initial reconstruction network optimizes the entire reconstruction image, not just individual independent image blocks, thereby leveraging intra-block and inter-block information to better optimize the reconstruction. The sampling network and initial reconstruction network are depicted in the Figure 2.

2.2.3. Deep Reconstruction

The core of the entire SALSA-Net is deep reconstruction, which is composed of multiple cascaded modules. Each module consists of a gradient update module (GUM), a thresholding denoising module (TDM), and an auxiliary update module (AUM).
1.
The thresholding denoising module v is designed to map the first iteration of the traditional SALSA algorithm onto the deep network architecture, aiming to eliminate the artifact noise in x + m using convolutional neural networks and the thresholding functions. The process can be expressed as:
v ^ = x + m + F ( x + m ) ,
where F ( · ) is designed as a sequence of convolutional operations, which are specifically defined as:
F ( · ) = B ( H ˜ ( L soft ( H ( C ( · ) ) ) ) .
The implementation of this module is illustrated in Figure 3, where C ( · ) is a one-shot convolutional operation that performs a linear transformation to increase the dimensionality of the input using 32 3 × 3 convolutional kernels; H ( · ) is designed to consist of two convolutional layers and a ReLU non-linear transformation layer to transform the output of C ( · ) into the desired domain and then perform denoising using a soft thresholding function L soft ( · ) . The output is then transformed back to the original domain using the transformation H ˜ ( · ) , satisfying H H ˜ = I . Finally, a series of convolutional operations B ( · ) are used to achieve dimensionality reduction and obtain the final output u. The reconstructed results of C ( · ) and B ( · ) are stacked with the previous image residual information to obtain the updated reconstruction results. Unlike ISTA-Net, our B ( · ) is a deep network that learns the sparse representation of the input image using multiple convolutional layers and ReLU activation functions. Thus, B ( · ) becomes a trainable module that can adapt to different image scenes and tasks to improve the CS reconstruction performance.
2.
The gradient update module is utilized in the SALSA algorithm to map the update process of x to the neural network. This module enables the learned M ( · ) from the sampling network to replace the sampling matrix A in the SALSA algorithm and the learned M ˜ ( · ) from the initial reconstruction network to replace A T . This approach eliminates the need for manual design of the sampling matrix in traditional algorithms and allows for sharing of the convolutional kernel parameters with those of the sampling and initial reconstruction, thereby improving network performance. Moreover, the module utilizes the network training parameter step size μ to avoid manual parameter tuning. The process can be expressed as:
x ^ = ( μ I + M ( M ˜ ( x ) ) ) 1 ( M ˜ ( y ) + μ ( v m ) ) .
3.
The auxiliary update module is a linear combination of the previous two modules. Its main purpose is to accelerate the convergence speed of the algorithm, enabling a faster search for the optimal solution. Moreover, this module utilizes the computed m value as the initial value for the next iteration, which is integrated into the iterative computation. The process can be expressed as:
m ^ = m v + x .

2.2.4. Parameter and Loss Function Design

1.
Parameters: The set of learnable parameters in our model comprises the transformation parameters M ( · ) , M ˜ ( · ) , C ( · ) , H ( · ) , H ˜ ( · ) , and B ( · ) . In addition, the step size μ and the shrinkage threshold β are also learnable, without the need for manual tuning. These parameters are shared across all steps of the reconstruction stage and are part of the deep neural network.
To ensure the correct convergence of the parameters μ and β , we introduce some constraints in the following manner:
β k = φ ( a 1 k + b 1 ) , a 1 < 0 , μ k = φ ( a 2 k + b 2 ) , a 2 < 0 .
Considering the decreasing noise variance during the iterative process, the shrinkage threshold is gradually decreased, and the step size should decrease smoothly during iterations. We enforce this constraint using the soft thresholding function φ ( · ) = ln ( 1 + exp ( · ) ) . Since the network is fully shared and { a 1 , a 2 , b 1 , b 2 } is independent of iterations, we can use a different number of iterations for image reconstruction, as described in Section 3.
2.
The design of the loss function: we define the original training set as { x i } i = 1 K b and the recovered images as { x i ˜ } i = 1 K p , where K b is the number of images in the training set and K p is the total number of stages in the reconstruction network.
To facilitate comparative experiments, we quantify the difference between the original and reconstructed images using mean squared error. Inspired by sparse auto-encoders and block-based image reconstruction, we aim to minimize the difference between the reconstructed and original images. Therefore, we design the loss functions to include three parts:
L total = L mse + λ 1 L sym + λ 2 L init ,
where
L mse = 1 K b K B 2 i = 1 K b x i ˜ x i F 2 ,
L sym = 1 K u i = 1 K b i = 1 K p H ˜ ( H ( u i ) ) u i F 2 ,
L init = 1 K b i = 1 K p M ˜ ( M ( x i ) ) x i F 2 .
The loss function of SALSA-Net consists of three parts: L mse aims to minimize the difference between the original and reconstructed images, where . F 2 represents the Frobenius norm of a matrix or tensor. L sym aims to ensure symmetry by making the inverse transform of H ˜ ( · ) as close as possible to H ( · ) , where u i = C ( x + m ) , K u is the number of elements in u i , and K B is the size of the image block. L init is the constraint imposed on the initial reconstruction. Regularization parameters λ 1 and λ 2 are set to 0.01 and 0.001 by default.

3. Results

3.1. Training Configuration

This study utilized the widely used Train91 dataset to train the models, extracting a total of 88,912 randomly cropped image patches of size 33 × 33 as the training set to ensure a fair comparison of experimental results. The performance of the proposed approach was assessed on three benchmark datasets: Set5 [55], Set11 [52], and BSD68 [56]. Set5 comprises 5 color images, Set11 comprises 11 grayscale images, and BSD68 comprises 68 natural color images.
This study trained the model using six different sampling rates, 10%, 20%, 25%, 30%, 40%, and 50%, and evaluated reconstruction quality using the peak signal-to-noise ratio (PSNR) metric. Higher PSNR values indicate better reconstructed image quality. All models were trained for 160 epochs with a batch size of 64, a learning rate of 0.0001, and an initial bias value of 0. The initial values of { a 1 , a 2 , b 1 , b 2 } are set to −0.4, −0.2, −2, and −1. The experiments were conducted on a platform equipped with a Quadro RTX 6000 GPU.

3.2. Analysis of Experimental Results

In this section, we compared our proposed method with several benchmark algorithms, including TVAL3, ReconNet, ISTA-Net+, AMP-Net, NL-CSNet, MAC-Net, and ISTA-Net++. TVAL3 is a model-based method, ReconNet is a classic deep network method, and ISTA-Net+, AMP-Net, NL-CSNet, MAC-Net, and ISTA-Net++ are all deep unrolling methods. We made uniform modifications to all methods to address the block-based image problem.
Firstly, to investigate the effect of the number of cascades in the network model on the reconstruction quality, we conducted a comparative analysis of our proposed method with ISTA-Net and ISTA-Net+ on the Set11 dataset with a 25% sampling rate. As depicted in Figure 4, the reconstruction quality of all methods improves as the number of cascades increases. Moreover, our proposed method outperforms ISTA-Net and ISTA-Net+ in terms of PSNR when the number of cascades is 7, 9, 11, 13, and 15. When the number of cascades exceeds 9, the improvement in reconstruction quality becomes insignificant. Therefore, to strike a balance between computational complexity and restoration performance, we set the number of cascades to 9 in the experiments conducted in this section.
We conducted a performance comparison of three loss functions, namely, L mse , L sym , and L init , during the iterative process of SALSA-Net on the Set11 dataset with a sampling rate of 25%. Figure 5 illustrates that SALSA-Net demonstrates a consistent and rapid convergence towards zero on all three loss functions, indicating its excellent convergence properties.
To validate the convergence of our proposed method, experiments were conducted on the Set11 dataset with a sampling rate of 25% and N = 9. The proposed method was compared with ISTA-Net and ISTA-Net+ and the experimental results are presented in Figure 6. The experimental findings indicate that the proposed method outperforms ISTA-Net and ISTA-Net+ at different epochs. Notably, the proposed method exhibits superior performance after 40 epochs, while the three methods show a slow improvement in the subsequent epochs and slightly decrease after around 160 epochs. To balance computational complexity and reconstruction performance, we set the number of epochs to 160 in the subsequent experiments.
Table 1 presents the average PSNR of eight different algorithm models on the Set11 dataset under different sampling rates of 10%, 20%, 30%, 40%, and 50%. Bold numbers highlight the best reconstruction quality results at each sampling rate. The findings reveal that TVAL3 performs poorly at low sampling rates, while ReconNet exhibits poor performance at high sampling rates. The deep unrolling models (ISTA-Net+, AMP-Net+, NL-CSNet, MAC-Net, and ISTA-Net++) outperform classical compressive sensing models (TVAL3 and ReconNet) in terms of reconstruction results. Although the proposed SALSA-Net has slightly lower average PSNR than AMP-Net and ISTA-Net++ at sampling rates of 10% and 25%, respectively, it outperforms other reconstruction methods as the sampling rate increases, indicating its effectiveness in CS image reconstruction.
Table 2 and Table 3 present the comparison of the experimental results of SALSA-Net and other models on the BSD68 and Set5 datasets, respectively, to evaluate the generalization ability of SALSA-Net. The tables highlight the best results in bold. The findings of both tables indicate that SALSA-Net outperforms other algorithms at low sampling rates, exhibiting the highest average PSNR. Specifically, in BSD68, SALSA-Net achieves an average improvement of 0.67 dB over ISTA-Net++, 1.51 dB over MAC-Net, and 4.14 dB over ReconNet. In Set5, SALSA-Net achieves an average improvement of 0.13dB over AMP-Net, 0.67 dB over ISTA-Net++, and 6.37 dB over TVAL3. Moreover, SALSA-Net exhibits better performance than NL-CSNet, ISTA-Net+, and other algorithms at various sampling rates. These results demonstrate the good generalization ability and high performance levels of SALSA-Net.
This study conducted a comparative analysis of the proposed SALSA-Net method against four other methods using partial reconstruction images at a 25% sampling rate. The partial reconstruction results of Set11, BSD68, and Set5 are presented in Figure 7, Figure 8 and Figure 9, respectively. The reconstruction results of ReconNet were found to be blurry, while the other methods achieved effective reconstruction to a certain extent. Furthermore, to assess the scalability of SALSA-Net, a set of medical brain images was used to evaluate its performance on the CS-MRI reconstruction problem. As shown in Figure 10, the proposed method achieves CS-MRI reconstruction. The experimental results indicate that the SALSA-Net method can reconstruct texture structures clearly, exhibiting clear visual performance and signal reconstruction accuracy.

4. Conclusions and Future Work

In this paper, we proposed SALSA-Net, a deep unrolling network designed to address the compressed sensing problem with images. By combining the interpretability of SALSA with the powerful learning ability of deep networks, SALSA-Net incorporates learnable sampling and residual modules to achieve superior denoising and detail restoration performance. Extensive experiments on large datasets demonstrate the effectiveness of SALSA-Net’s sampling training and reconstruction strategy, which outperforms other state-of-the-art algorithms. In the future, we intend to introduce cross-area modules to further enhance the model’s performance and increase its flexibility. To expedite the study of deep unrolling networks, we will release the source codes and dataset of this paper to the public at https://github.com/songhp/SALSANet (accessed on 23 May 2023).

Author Contributions

Conceptualization, H.S. and Q.D.; methodology, H.S. and Q.D.; software, Q.D. and J.G.; supervision, H.S. and Y.L.; writing—original draft, H.S. and Q.D.; writing—review and editing, H.S., H.M. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (Grant Nos. 62172193, 62272051) and the General Project of Science and Technology Plan of Beijing Municipal Commission of Education (No. KM201910009014).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the Editor-in-Chief, editors, and anonymous reviewers for their valuable reviews.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  2. Candes, E.; Romberg, J.; Tao, T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar] [CrossRef]
  3. Candes, E.; Tao, T. Near-optimal signal recovery from random projections: Universal encoding strategies? IEEE Trans. Inf. Theory 2006, 52, 5406–5425. [Google Scholar] [CrossRef]
  4. Wang, L.; Lu, K.; Liu, P. Compressed sensing of a remote sensing image based on the priors of the reference image. IEEE Geosci. Remote Sens. Lett. 2015, 12, 736–740. [Google Scholar] [CrossRef]
  5. Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Sun, T.; Kelly, K.F.; Baraniuk, R.G. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef]
  6. Wen, B.; Ravishankar, S.; Pfister, L.; Bresler, Y. Transform learning for magnetic resonance image reconstruction: From model-based learning to building neural networks. IEEE Signal Process. Mag. 2020, 37, 41–53. [Google Scholar] [CrossRef]
  7. Li, S.; Da Xu, L.; Wang, X. Compressed sensing signal and data acquisition in wireless sensor networks and internet of things. IEEE Trans. Ind. Inform. 2012, 9, 2177–2186. [Google Scholar] [CrossRef]
  8. Potter, L.; Ertin, E.; Parker, J.; Cetin, M. Sparsity and compressed sensing in radar imaging. Proc. IEEE 2010, 98, 1006–1020. [Google Scholar] [CrossRef]
  9. Duarte, M.F.; Baraniuk, R.G. Spectral compressive sensing. Appl. Comput. Harmon. Anal. 2013, 35, 111–129. [Google Scholar] [CrossRef]
  10. Wright, J.; Ma, Y.; Mairal, J.; Sapiro, G.; Huang, T.; Yan, S. Sparse representation for computer vision and pattern recognition. Proc. IEEE 2010, 98, 1031–1044. [Google Scholar] [CrossRef]
  11. Gan, L. Block Compressed Sensing of Natural Images. In Proceedings of the International Conference on Digital Signal Processing (ICDSP), Cardiff, UK, 1–4 July 2007. [Google Scholar]
  12. Eftekhari, A.; Yap, H.L.; Rozell, C.J.; Wakin, M.B. The restricted isometry property for random block diagonal matrices. Appl. Comput. Harmon. Anal. 2015, 38, 1–31. [Google Scholar] [CrossRef]
  13. Bruckstein, A.M.; Donoho, D.L.; Elad, M. From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images. Siam Rev. 2009, 51, 34–81. [Google Scholar] [CrossRef]
  14. Tropp, J.; Wright, S. Computational methods for sparse solution of linear inverse problems. Proc. IEEE 2010, 98, 948–958. [Google Scholar] [CrossRef]
  15. Wang, Y.; Meng, D.; Yuan, M. Sparse recovery: From vectors to tensors. Natl. Sci. Rev. 2018, 5, 756–767. [Google Scholar] [CrossRef]
  16. Song, H.; Ai, Z.; Lai, Y.; Meng, H.; Mao, Q. Sparse signal reconstruction via generalized two-stage thresholding. Sci. China Inf. Sci. 2022, 65, 139303. [Google Scholar] [CrossRef]
  17. Donoho, D.; Maleki, A.; Montanari, A. Message-passing algorithms for compressed sensing. Proc. Natl. Acad. Sci. USA 2009, 106, 18914–18919. [Google Scholar] [CrossRef]
  18. Parikh, N.; Boyd, S. Proximal Algorithms. Found. Trends Optim. 2014, 1, 127–239. [Google Scholar] [CrossRef]
  19. Li, C.; Yin, W.; Jiang, H.; Zhang, Y. An efficient augmented Lagrangian method with applications to total variation minimization. Comput. Optim. Appl. 2013, 56, 507–530. [Google Scholar] [CrossRef]
  20. Needell, D.; Tropp, J.A. CoSaMP: Iterative signal recovery from incomplete and inaccurate samples. Appl. Comput. Harmon. Anal. 2009, 26, 301–321. [Google Scholar] [CrossRef]
  21. Dai, W.; Milenkovic, O. Subspace pursuit for compressive sensing signal reconstruction. IEEE Trans. Inf. Theory 2009, 55, 2230–2249. [Google Scholar] [CrossRef]
  22. Blumensath, T.; Davies, M.E. Iterative hard thresholding for compressed sensing. Appl. Comput. Harmon. Anal. 2009, 27, 265–274. [Google Scholar] [CrossRef]
  23. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
  24. Becker, S.; Bobin, J.; Candes, E. NESTA: A fast and accurate first-order method for sparse recovery. SIAM J. Imaging Sci. 2011, 4, 1–39. [Google Scholar] [CrossRef]
  25. Afonso, M.; Bioucas-Dias, J.; Figueiredo, M. Fast image recovery using variable splitting and constrained optimization. IEEE Trans. Image Process. 2010, 19, 2345–2356. [Google Scholar] [CrossRef]
  26. Afonso, M.; Bioucas-Dias, J.; Figueiredo, M. An augmented Lagrangian approach to the constrained optimization formulation of imaging inverse problems. IEEE Trans. Image Process. 2010, 20, 681–695. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, J.; Chen, B.; Xiong, R.; Zhang, Y. Physics-inspired compressive sensing: Beyond deep unrolling. IEEE Signal Process. Mag. 2023, 40, 58–72. [Google Scholar] [CrossRef]
  28. Monga, V.; Li, Y.; Eldar, Y.C. Algorithm unrolling: Interpretable, efficient deep learning for signal and image processing. IEEE Signal Process. Mag. 2021, 38, 18–44. [Google Scholar] [CrossRef]
  29. Chen, T.; Chen, X.; Chen, W.; Heaton, H.; Liu, J.; Wang, Z.; Yin, W. Learning to optimize: A primer and a benchmark. J. Mach. Learn. Res. 2022, 23, 1–59. [Google Scholar]
  30. Mousavi, A.; Patel, A.B.; Baraniuk, R.G. A deep learning approach to structured signal recovery. In Proceedings of the 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 29 September–2 October 2015; pp. 1336–1343. [Google Scholar]
  31. Kulkarni, K.; Lohit, S.; Turaga, P.; Kerviche, R.; Ashok, A. Reconnet: Non-iterative reconstruction of images from compressively sensed measurements. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 449–458. [Google Scholar]
  32. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  33. Yao, H.; Dai, F.; Zhang, S.; Zhang, Y.; Tian, Q.; Xu, C. DR2-Net: Deep residual reconstruction network for image compressive sensing. Neurocomputing 2019, 359, 483–493. [Google Scholar] [CrossRef]
  34. Shi, W.; Jiang, F.; Zhang, S.; Zhao, D. Deep networks for compressed image sensing. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Hong Kong, China, 10–14 July 2017; pp. 877–882. [Google Scholar]
  35. Sun, Y.; Yang, Y.; Liu, Q.; Chen, J.; Yuan, X.T.; Guo, G. Learning non-locally regularized compressed sensing network with half-quadratic splitting. IEEE Trans. Multimed. 2020, 22, 3236–3248. [Google Scholar] [CrossRef]
  36. Chen, B.; Zhang, J. Content-aware scalable deep compressed sensing. IEEE Trans. Image Process. 2022, 31, 5412–5426. [Google Scholar] [CrossRef]
  37. Zhou, S.; He, Y.; Liu, Y.; Li, C.; Zhang, J. Multi-Channel Deep Networks for Block-Based Image Compressive Sensing. IEEE Trans. Multimed. 2021, 23, 2627–2640. [Google Scholar] [CrossRef]
  38. You, D.; Zhang, J.; Xie, J.; Chen, B.; Ma, S. Coast: Controllable arbitrary-sampling network for compressive sensing. IEEE Trans. Image Process. 2021, 30, 6066–6080. [Google Scholar] [CrossRef]
  39. Gregor, K.; LeCun, Y. Learning fast approximations of sparse coding. In Proceedings of the International Conference on Machine Learning (ICML), Haifa, Israel, 21–24 June 2010; pp. 399–406. [Google Scholar]
  40. Chen, X.; Liu, J.; Wang, Z.; Yin, W. Theoretical linear convergence of unfolded ISTA and its practical weights and thresholds. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montreal, QB, Canada, 2–8 December 2018; pp. 9061–9071. [Google Scholar]
  41. Liu, J.; Chen, X.; Wang, Z.; Yin, W. ALISTA: Analytic weights are as good as learned weights in LISTA. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  42. Wu, K.; Guo, Y.; Li, Z.; Zhang, C. Sparse coding with gated learned ISTA. In Proceedings of the International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 26 April–May 1 2020. [Google Scholar]
  43. Li, Y.; Kong, L.; Shang, F.; Liu, Y.; Liu, H.; Lin, Z. Learned extragradient ISTA with interpretable residual structures for sparse coding. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Online, 2–9 February 2021; Volume 35, pp. 8501–8509. [Google Scholar]
  44. Aberdam, A.; Golts, A.; Elad, M. Ada-LISTA: Learned Solvers Adaptive to Varying Models. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9222–9235. [Google Scholar] [CrossRef]
  45. Zheng, Z.; Dai, W.; Xue, D.; Li, C.; Zou, J.; Xiong, H. Hybrid ISTA: Unfolding ISTA with convergence guarantees using free-form deep neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 43, 3226–3242. [Google Scholar] [CrossRef] [PubMed]
  46. Zhang, J.; Ghanem, B. ISTA-Net: Interpretable optimization-inspired deep network for image compressive sensing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 1828–1837. [Google Scholar]
  47. You, D.; Xie, J.; Zhang, J. ISTA-NET++: Flexible deep unfolding network for compressive sensing. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021. [Google Scholar]
  48. Xiang, J.; Dong, Y.; Yang, Y. FISTA-Net: Learning a Fast Iterative Shrinkage Thresholding Network for Inverse Problems in Imaging. IEEE Trans. Med. Imaging 2021, 40, 1329–1339. [Google Scholar] [CrossRef]
  49. Yang, Y.; Sun, J.; Li, H.; Xu, Z. ADMM-CSNet: A deep learning approach for image compressive sensing. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 521–538. [Google Scholar] [CrossRef]
  50. Liu, J.; Sun, Y.; Eldeniz, C.; Gan, W.; An, H.; Kamilov, U.S. RARE: Image reconstruction using deep priors learned without groundtruth. IEEE J. Sel. Top. Signal Process. 2020, 14, 1088–1099. [Google Scholar] [CrossRef]
  51. Zhang, Z.; Liu, Y.; Liu, J.; Wen, F.; Zhu, C. AMP-Net: Denoising-based deep unfolding for compressive image sensing. IEEE Trans. Image Process. 2021, 30, 1487–1500. [Google Scholar] [CrossRef]
  52. Chen, J.; Sun, Y.; Liu, Q.; Huang, R. Learning memory augmented cascading network for compressed sensing of images. In Proceedings of the European Conference Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 513–529. [Google Scholar]
  53. Ye, D.; Ni, Z.; Wang, H.; Zhang, J.; Wang, S.; Kwong, S. CSformer: Bridging Convolution and Transformer for Compressive Sensing. arXiv 2021, arXiv:2112.15299. [Google Scholar] [CrossRef] [PubMed]
  54. Shen, M.; Gan, H.; Ning, C.; Hua, Y.; Zhang, T. TransCS: A Transformer-Based Hybrid Architecture for Image Compressed Sensing. IEEE Trans. Image Process. 2022, 31, 6991–7005. [Google Scholar] [CrossRef] [PubMed]
  55. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the British Machine Vision Conference (BMVC), Surrey, UK, 3–7 September 2012; pp. 1–10. [Google Scholar]
  56. Sun, Y.; Chen, J.; Liu, Q.; Liu, B.; Guo, G. Dual-path attention network for compressed sensing image reconstruction. IEEE Trans. Image Process. 2020, 29, 9482–9495. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Illustration of SALSA-Net framework.
Figure 1. Illustration of SALSA-Net framework.
Sensors 23 05142 g001
Figure 2. Illustration of the sampling and initial reconstruction process.
Figure 2. Illustration of the sampling and initial reconstruction process.
Sensors 23 05142 g002
Figure 3. Illustration of the thresholding denoising module.
Figure 3. Illustration of the thresholding denoising module.
Sensors 23 05142 g003
Figure 4. Comparison of average PSNR among ISTA-Net, ISTA-Net+, and SALSA-Net with different phases.
Figure 4. Comparison of average PSNR among ISTA-Net, ISTA-Net+, and SALSA-Net with different phases.
Sensors 23 05142 g004
Figure 5. Updated diagrams of L mse , L sym , and L init during the iteration process.
Figure 5. Updated diagrams of L mse , L sym , and L init during the iteration process.
Sensors 23 05142 g005
Figure 6. Comparisons of the average PSNR among ISTA-Net, ISTA-Net+, and SALSA-Net with different epochs.
Figure 6. Comparisons of the average PSNR among ISTA-Net, ISTA-Net+, and SALSA-Net with different epochs.
Sensors 23 05142 g006
Figure 7. Visual comparisons of original image and reconstructions by ReconNet, ISTA-Net+, AMP-Net, and SALSA-Net on Set11 with a sampling rate of 25% after 160 epochs, arranged from left to right.
Figure 7. Visual comparisons of original image and reconstructions by ReconNet, ISTA-Net+, AMP-Net, and SALSA-Net on Set11 with a sampling rate of 25% after 160 epochs, arranged from left to right.
Sensors 23 05142 g007
Figure 8. Visual comparisons of original image and reconstructions by ReconNet, ISTA-Net+, AMP-Net, and SALSA-Net on BSD68 with a sampling rate of 25% after 160 epochs, arranged from left to right.
Figure 8. Visual comparisons of original image and reconstructions by ReconNet, ISTA-Net+, AMP-Net, and SALSA-Net on BSD68 with a sampling rate of 25% after 160 epochs, arranged from left to right.
Sensors 23 05142 g008
Figure 9. Visual comparisons of original image and reconstructions by ReconNet, ISTA-Net+, AMP-Net, and SALSA-Net on Set5 with a sampling rate of 25% after 160 epochs, arranged from left to right.
Figure 9. Visual comparisons of original image and reconstructions by ReconNet, ISTA-Net+, AMP-Net, and SALSA-Net on Set5 with a sampling rate of 25% after 160 epochs, arranged from left to right.
Sensors 23 05142 g009
Figure 10. Visual comparisons of original brain images (top) and reconstructions by SALSA-Net (bottom) with a sampling rate of 25%.
Figure 10. Visual comparisons of original brain images (top) and reconstructions by SALSA-Net (bottom) with a sampling rate of 25%.
Sensors 23 05142 g010
Table 1. PSNR performance comparisons between SALSA-Net and seven other algorithms on Set11.
Table 1. PSNR performance comparisons between SALSA-Net and seven other algorithms on Set11.
Methods10%20%30%40%50%Avg
PSNR(dB)
TVAL322.9927.9229.2331.4633.5529.03
ReconNet24.1125.5228.7130.7731.5928.14
ISTA-Net+26.5430.6733.7236.0138.0132.99
AMP-Net28.6531.0532.9135.3137.4533.07
NL-CSNet28.1131.4333.8235.6037.1333.22
MAC-Net27.8331.5433.8236.1037.7833.41
ISTA-Net++28.3432.3334.8636.5138.7334.15
SALSA-Net28.4932.2535.2036.9739.3934.46
Table 2. PSNR performance comparisons between SALSA-Net and seven other algorithms on BSD68.
Table 2. PSNR performance comparisons between SALSA-Net and seven other algorithms on BSD68.
Methods10%20%30%40%50%Avg
PSNR(dB)
TVAL319.2621.2522.3425.3929.5923.57
ReconNet23.7625.2927.4028.5830.6427.13
ISTA-Net+25.2927.2330.0332.2333.5629.67
AMP-Net25.3227.3730.5632.1132.7829.63
NL-CSNet25.1827.5329.6131.3232.4829.22
MAC-Net25.3428.4330.1131.3733.5829.76
ISTA-Net++26.0128.5630.9432.7234.9230.63
SALSA-Net26.9629.1231.5531.5535.6031.27
Table 3. PSNR performance comparisons between SALSA-Net and six other algorithms on Set5.
Table 3. PSNR performance comparisons between SALSA-Net and six other algorithms on Set5.
Methods10%20%30%40%50%Avg
PSNR(dB)
TVAL327.1230.3532.4934.7736.6332.27
ReconNet26.9929.3431.4433.6235.4131.36
ISTA-Net+28.8432.6235.4738.4540.1535.10
AMP-Net33.3136.9239.1441.1242.0738.51
MAC-Net32.5436.0638.4840.3742.1137.91
ISTA-Net++33.0436.4138.8639.9241.6437.97
SALSA-Net33.7836.6339.2141.3542.2438.64
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, H.; Ding, Q.; Gong, J.; Meng, H.; Lai, Y. SALSA-Net: Explainable Deep Unrolling Networks for Compressed Sensing. Sensors 2023, 23, 5142. https://doi.org/10.3390/s23115142

AMA Style

Song H, Ding Q, Gong J, Meng H, Lai Y. SALSA-Net: Explainable Deep Unrolling Networks for Compressed Sensing. Sensors. 2023; 23(11):5142. https://doi.org/10.3390/s23115142

Chicago/Turabian Style

Song, Heping, Qifeng Ding, Jingyao Gong, Hongying Meng, and Yuping Lai. 2023. "SALSA-Net: Explainable Deep Unrolling Networks for Compressed Sensing" Sensors 23, no. 11: 5142. https://doi.org/10.3390/s23115142

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop