Next Article in Journal
ECG Approximate Entropy in the Elderly during Cycling Exercise
Next Article in Special Issue
Bed-Exit Behavior Recognition for Real-Time Images within Limited Range
Previous Article in Journal
EEG Emotion Classification Network Based on Attention Fusion of Multi-Channel Band Features
Previous Article in Special Issue
Sentiment Analysis: An ERNIE-BiLSTM Approach to Bullet Screen Comments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RCA-LF: Dense Light Field Reconstruction Using Residual Channel Attention Networks

1
School of Information and Communication Engineering, College of Electrical and Computer Engineering, Chungbuk National University, Cheongju 28644, Korea
2
Electrical Engineering Department, Faculty of Engineering, Assiut University, Assiut 71515, Egypt
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(14), 5254; https://doi.org/10.3390/s22145254
Submission received: 24 May 2022 / Revised: 9 July 2022 / Accepted: 11 July 2022 / Published: 14 July 2022
(This article belongs to the Special Issue AI Multimedia Applications)

Abstract

:
Dense multi-view image reconstruction has played an active role in research for a long time and interest has recently increased. Multi-view images can solve many problems and enhance the efficiency of many applications. This paper presents a more specific solution for reconstructing high-density light field (LF) images. We present this solution for images captured by Lytro Illum cameras to solve the implicit problem related to the discrepancy between angular and spatial resolution resulting from poor sensor resolution. We introduce the residual channel attention light field (RCA-LF) structure to solve different LF reconstruction tasks. In our approach, view images are grouped in one stack where epipolar information is available. We use 2D convolution layers to process and extract features from the stacked view images. Our method adopts the channel attention mechanism to learn the relation between different views and give higher weight to the most important features, restoring more texture details. Finally, experimental results indicate that the proposed model outperforms earlier state-of-the-art methods for visual and numerical evaluation.

1. Introduction

Light fields (LF) record 3D scenes into uniform and dense image samples. These images contain spatial and angular information about the 3D scenes. As a result, many applications have developed and benefited greatly from this huge amount of information, such as de-occlusion [1,2], depth-sensing [3,4,5], saliency detection [6], and salient object detection [7]. In addition, LF could be promising to ease other applications such as the fruit-picking robot, where a robot traverses a whole field and harvests on its own [8,9]. LF images are caught using portable cameras or camera arrays in most situations. In order to use array cameras, several cameras are required, which is an expensive and laborious process [10]. A practical solution for capturing LF images with portable cameras can be provided by inserting a microlens array in front of the image sensor [11,12]. Despite the advantages of this solution, it comes with a major drawback: poor sensor resolution. Therefore, obtaining LF images with high spatial and angular resolution is difficult.
Recently, several learning-based approaches that considerably enhance the performance of LF reconstruction have been presented. The LF reconstruction challenge reconstructs dense LF images from sparse input views. Previous approaches using the convolutional neural network (CNN) without depth estimation [13,14] can only handle LFs with a small baseline. They explore the connection between the angular and spatial domains but fail to use the epipolar information fully.
Some approaches [15,16] estimate depth maps and warp views to investigate relationships between views. However, the wrongness of the calculated depth map greatly affects how the LF reconstruction turns out. There is another approach to mitigate the effect of limited sensor resolution through LF super-resolution [17,18,19], but this is outside this research’s interest.
This article presents a unique learning-based methodology for rapidly reconstructing a densely sampled LF from a very sparsely sampled LF. Computationally efficient convolutions realize our end-to-end CNN model to understand spatial-angular relationships deeply. We up-sample the sparsely input LF to the required angular size using the bicubic interpolation in the preprocessing stage. The RCA-LF is then deployed to leverage the inherent LF structure in the up-sampled LF images. Notably, our method does not need disparity warping or intensive computations. In addition, it reconstructs a whole LF in a single forward pass. Specifically, we introduce the residual channel attention light field (RCA-LF) structure to solve different LF reconstruction tasks. In our approach, view images are grouped in one stack where epipolar information is available. We use 2D convolution layers to process and extract features from the stacked view images. Our method adopts the channel attention mechanism to learn the relation between different views and give higher weights to the most important features, restoring more texture details.
We propose a new way to process the multi-channel input, which comes from 2D convolution instead of 3D convolution. Two-dimensional convolution takes a single slice as an input and fails to leverage context from adjacent slices. Conversely, 3D convolution overcomes this issue by leveraging the slice context with 3D convolutional kernels, resulting in enhanced performance. However, 3D convolutions have a limited range depending on the kernel size (3 × 3 × 3 kernels can leverage depth information using only three consecutive slices).
In our proposed method, the input has a size of (B, H, W, 49) for the 3 × 3 to 7 × 7 reconstruction task where the 49 represents the number of input channels. For 2D convolution, the number of filters equals (filter_height × filter_width × in_channels × out_channels). Consequently, every output channel is a function of all input channels at each convolution. Adopting this method can fix the limited range issue of the 3D convolution and provide better quality.
The number of input channels is 49 for the 3 × 3 to 7 × 7 reconstruction task. Still, we extract more features on the subsequent convolution layers, meaning more interactions can be identified between the extracted features of input images, restoring more information and details. Because some of the extracted features might contain useless or redundant information, the channel attention mechanism rescales (gives different weights) for these extracted features depending on the information content.
We can summarize the contributions of this article as follows: (1) We adopt a channel attention mechanism to reconstruct LF images. (2) Our method increases the interaction between different LF images by processing LF images as input–output channels of 2D convolutions. (3) We design the RCA-LF to increase the interactions between input–output channels (parallel processing) and decrease the number of blocks (serial processing); hence, it can reconstruct LF images accurately and fast.

2. Related Work

2.1. LF Representation

A wealth of information about the surrounding 3D space is revealed by LF imaging, contrary to traditional imaging methods. The Plenoptic function was initially described using seven variables that determine the view from any possible angle, for all wavelengths of light and at any time, as P = P ( θ , φ , λ , t , V x , V y , V z ) [20]. It was then simplified to a 4D description with the intersections of light rays with two planes L = L ( u , v , x , y ) , where (u, v) and (x, y) denote the points of intersection with the first and second planes, respectively, as shown in Figure 1 [21].

2.2. LF Reconstruction

Many LF reconstruction approaches have been presented. These approaches are classified into three types: traditional, deep learning depth-based, and deep learning non-depth-based approaches.

2.2.1. Traditional Approaches

Wanner and Goldluecke improved the spatial and angular resolutions using the Epipolar Plane Image (EPI) for depth map estimation [22]. However, this variational framework has flaws since the input views only assess the disparity. Another approach was proposed to utilize the Gaussian mixture model for LF denoising, super-resolution, and refocusing [23]. In this approach, the patch prior was designed using the disparity pattern. However, their approach is vulnerable to low-quality LF images. Pujades et al. [24] proposed a novel cost function optimized by a Bayesian formulation to estimate the depth and reconstruct novel views. Chaurasia et al. [25] proposed a novel image-based rendering using superpixels to preserve depth discontinuities. The warped views are blended using a camera and depth information. Zhang et al. [26] proposed an interactive system adopting patch-based methods for LF editing. This technique models the collected images as overlapping layers with varying depths and uses back-to-front layered synthesis. Vagharshakyan et al. [27] utilized the EPIs in the shearlet domain to reconstruct dense images using large baseline-rectified images. Their method provided good results for non-Lambertian scenes of semi-transparent objects.

2.2.2. Deep Learning Depth-Based Approaches

Kalantari et al. [15] suggested decomposing the reconstruction process into disparity and color estimates independently evaluated by the relevant CNN network. Due to their separate reconstruction, connections between novel LF images were overlooked. Another approach was proposed to speed up Kalantari’s method using a predefined CNN [28]. In addition, they proposed the estimation of two disparity maps to provide more accurate results. Shi et al. [16] used two reconstruction modules: pixel reconstruction to handle the occlusions explicitly, and feature reconstruction for high frequencies. However, this method was limited by the need for depth maps. In contrast to the previous methods designed for images with a small baseline, Jin et al. [29] designed a model for images with a large baseline. A CNN was employed to estimate depth maps to wrap input views, and these views were then blended using a SAS CNN [30]. Because the quality of synthesized views is dependent on the accuracy of estimated depth maps, unwanted artifacts often emerge in synthesized views.

2.2.3. Deep Learning Non-Depth-Based Approaches

Most of these methods extract information from EPIs for the reconstruction process. Wu et al. [31] divided the process into low-frequency restoration after a blur operation and high-frequency restoration by inverting the blur operation. However, they did not use the epipolar information efficiently, as they only extracted the EPIs in one direction. Using a CNN, Wu et al. [32] applied a shearing operation to input EPIs to eliminate the effect of significant disparities. Then, they employed a CNN to learn a fusion score. In this method, the authors misused the angular information by using EPIs horizontally or vertically for the reconstruction. In addition, they reconstructed rows and then columns hierarchically, leading to reconstruction error accumulation. Meng et al. [33] proposed an HDDRNet for LF spatial and angular super-resolution employing a high dimensional CNN. Although they used the provided angular information efficiently, employing the 4D convolutions, this was at the expense of model complexity. Mildenhall et al. [34] proposed reconstructing multi-plane images from input views and then blending them to reconstruct novel views. Wang et al. [35] used EPI and EPI stacking to create a pseudo-4D CNN. They used EPI structure-preserving loss to increase reconstruction quality. They wasted angular data by only using horizontal or vertical EPI stacks. Hu et al. [14] proposed LF reconstruction with hierarchical feature fusion. SAS layers were employed to extract features from 4D LF images, while the U-Net structure was adopted to generate both semantic and local feature representation. They integrated these two structures and proposed a U-SAS module to enable the extraction of spatial features and the correlation of SAIs. In addition, they adopted an enlarged patch size when training for the integrated information of objects. Liu et al. [36] proposed to extract EPI information in a horizontal, vertical, and angular manner to reconstruct LF images. However, each branch was processed alone, which affected the final quality. Zhang et al. [37] reconstructed LF images employing 2D and 3D CNNs on horizontal and vertical EPIs. However, they neglected the angular LF information, slightly affecting the final reconstruction quality. Salem et al. [38] mapped the LF reconstruction problem from the 4D into the 2D domain by transforming the 4D LF into a 2D raw LF image to ease the reconstruction. They provided satisfactory reconstruction quality using a model inspired by the RCAN [39,40]. Still, they used a heavy model, which affected the reconstruction time.

3. Methodology

3.1. Problem Formulation

We can consider the LF images as a 2D array of view images, as shown in Figure 2a. These images have (H, W) and (U, V) spatial and angular resolutions. Our goal is to reconstruct dense LF images from their sparse input counterparts. Assume L L R R H × W × u × v represents the sparse input views with angular resolution (u, v). Using the LR input, our RCA-LF network can reconstruct a dense output L H R R H × W × U × V with (U, V) angular resolution. Before applying the LR input images, we up-sample the sparse input EPIs to the required output size utilizing the Bicubic interpolation to generate L ˜ L R R H × W × U × V . The last step before applying the LR input to the network is to rearrange it from the 4D representation L ˜ L R R H × W × U × V into the 3D representation L ˜ L R R H × W × U V , as shown in Figure 2b. We reconstruct the 3D L ˜ L R by stacking the view images in row-major order as indicated by the blue line in Figure 2a.

3.2. Network Architecture

We designed our network similarly to the RCAN network [39]. In terms of functionality, our model can be divided into primary feature extraction, deep feature extraction, and final output restoration, as shown in Figure 3a. The primary feature extraction is implemented using two convolutional layers (Conv). Each Conv is followed by a long skip connection to bypass the low-frequency components to the output part, allowing the network to concentrate on high-frequency component extraction. The deep feature extraction is implemented using ten residual channel attention blocks (RCAB), as shown in Figure 3b. The final part is implemented by summing the primary extracted features with the deep extracted features to reconstruct the final output.
This is unlike the RCAN method, in which the input is a single-channel input. Then, channel-wise features are extracted from the input to be processed through the network. The input in our method is a stack of U × V images (multi-channel input), where (U, V) is the angular resolution. Then, more channel-wise features are extracted with the extraction ratio e to be e × U × V . The RCAB is the main component of our network, as the RCA-LF consists of ten RCABs. The RCAB is a residual block (RB) with an integrated channel attention mechanism (CA). The first part of the RCAB, RB, is built by cascading two Conv layers with an activation function (ReLU) with a skip connection.
The CA is adopted to allow the network to treat the extracted channel-wise features unequally and concentrate on the crucial features. A global average pooling is used to shrink the intermediate C feature map of size H × W into 1 × 1 to obtain the initial channel-wise statistics to determine which channels are more important. These channel statistics may be considered a collection of local descriptors to express the full-view stack [41]. A Conv then down-samples these initial statistics with a reduction ratio of r. A Conv up-samples these statistics with the same reduction ratio after being activated by ReLU, as shown in Figure 3 in [39]. Finally, a gate mechanism is applied to learn the nonlinear interactions between channels and the non-exclusive mutual relationship. The gate mechanism is applied with a sigmoid function to obtain the final channel-wise statistics.

3.3. Implementation Details

The luminance component is only used to train the RCA-LF network, while the EPIs of the chrominance components are up-sampled with the Bicubic interpolation. We trained our network to map the LR input images to the HR LF output images by minimizing the L1 loss and optimizing the Adam optimizer with its default parameters [42]. The L1 loss is defined as follows when a training set has N combinations of input and counterpart ground-truth pictures:
L 1 = 1 N i = 1 N | L H R i f ( L L R i ) |
f ( ) represents the function responsible for mapping the LR input into the HR output and is implemented by the RCA-LF network. All the Conv layers used were of size 3 × 3 with zero padding, except for the Conv layers used for the CA, which were of size 1 × 1 . Both the extraction ratio e and the reduction ratio were set to 8. We trained the network with patches of size 32 × 32 and a batch size of 128. We started the training with an initial learning rate of 10 4 and decreased it exponentially by 0.1 every 100 epochs while we trained the network for 150 epochs. We used 100 full LF images to train our network [15,43], using TensorFlow [44] on an NVIDIA GeForce RTX 3090 GPU. PSNR and SSIM were used as reconstruction quality assessment indicators.

4. Experiments and Discussion

We conducted comprehensive experiments to validate the effect of the proposed RAC_LF network. We compared the RCA_LF numerically and visually with state-of-the-art methods using real-world LF images. We used 30 LFs from the 30 scenes dataset [15], 31 LFs from the refractive and reflective surfaces dataset [43], and 43 LFs from the occlusions dataset [43]. The average PSNR and SSIM [45] over the reconstructed LF luminance were used for the numerical comparison. We compared the RCA_LF over two interpolation tasks (2 × 2–8 × 8 and 3 × 3–7 × 7) and two extrapolation tasks (2 × 2–8 × 8 extrapolations 1 and 2), as shown in Figure 4.
Table 1, Table 2, Table 3 and Table 4 present numerical data indicating the proposed approach’s effectiveness. Numerical comparisons are provided regarding peak-signal-to-noise ratio (PSNR) and the structural similarity index (SSIM) [45]. Figure 5, Figure 6, Figure 7 and Figure 8 show a visual contrast highlighting our model’s ability to recreate high-quality images with sharper edges around object boundaries, even in obscured areas and against complex backgrounds. However, we attribute the significant improvement in our model results to: (1) 3D representation (LF view stack), allowing the network to model and understand relations between different LFs; (2) the channel attention mechanism, which played an important role by allowing the network to concentrate on the crucial features.

4.1. Different Reconstruction Tasks

4.1.1. Task 3 × 3–7 × 7

Wu et al. [31] underutilized angular data, using EPIs in just one direction. Utilizing EPIs in horizontal and vertical dimensions subsequently yielded superior outcomes. Nonetheless, they hierarchically up-sampled LF, increasing error accumulation on the most recently reconstructed views. In addition, they proposed a second paradigm based on sheared EPIs [32]. In particular, low-angular-resolution EPIs were sheared before being up-sampled to the necessary angular resolution. The up-sampled EPIs with various shearing methods were fused by learning fusion scores using a CNN. Liu et al. [36] used angular information more effectively than earlier techniques, yet this was insufficient since they only employed one EPI stack in each direction. Zhang et al. [37] used micro-lens pictures and view image stacks to investigate further LF data. Salem et al. [38] used the raw LF representation to ease the reconstruction process. In addition, they initialized the input image using the nearest view initialization method. However, this method had a limitation for some reconstruction tasks. Additionally, it affected the quality of the final image.
Table 1. The proposed model’s numerical comparison (PSNR/SSIM) model to reconstruct 7 × 7 out of 3 × 3 views.
Table 1. The proposed model’s numerical comparison (PSNR/SSIM) model to reconstruct 7 × 7 out of 3 × 3 views.
DatasetWu [31]Wu [32]Liu [36]Zhang [37]Salem [38]Proposed
30 Scenes41.40/0.98043.592/0.98644.86/0.99145.68/0.99245.96/0.99146.41/0.992
Reflective42.19/0.97443.092/0.97744.31/0.98044.92/0.98245.45/0.98345.73/0.984
Occlusions37.25/0.92539.748/0.94840.16/0.95740.80/0.95541.21/0.95741.41/0.951
Average40.28/0.95942.14/0.97143.11/0.97643.80/0.97644.21/0.97744.51/0.976
Figure 5. Visual comparison of the proposed model to reconstruct 7 × 7 views out of 3 × 3.
Figure 5. Visual comparison of the proposed model to reconstruct 7 × 7 views out of 3 × 3.
Sensors 22 05254 g005

4.1.2. Task 2 × 2–8 × 8, Extrapolation 0

Both Kalantari et al. [15] and Shi et al. [16] generate new views by distorting the input views by their assessed disparity/depth. On the other hand, depth estimation and warping are challenging, particularly for LF pictures with a tiny depth difference, making it possible for images to be flawed and seem out of place. Due to Yeung et al.’s disregard for the connections between distinct views, their approach generates false shadows and ghosting artifacts at the borders of reconstructed views [46].

4.1.3. Task 2 × 2–8 × 8, Extrapolation 1, 2

Reconstructing 8 × 8 out of 2 × 2 views is a challenging task due to the sparseness of the input views. Yeung et al. [46] observed that the reconstruction quality of the center views is much worse than that of the views located near the input views. Because the center view is the farthest distance from any input views, inferring the details with greater accuracy presents the biggest problem. Therefore, they proposed different combinations of interpolation and extrapolation to reconstruct LF images. As a result, the average distance from all the novel views is shorter than before, increasing the reconstruction quality of the center views. Most available algorithms are optimized for interpolation tasks and cannot predict extrapolated views. That is why ghosting and artifacts often appear around thin structures and occluded regions. Extrapolation is more challenging than interpolation because certain portions of the reconstructed views are not present in the input. In addition, it cannot keep the slopes of the lines in the reconstructed EPIs the same. It is challenging to devise a method for dealing with different relationships between input and output views. However, the task becomes more feasible and efficient with our proposed approach.
Table 2. The proposed model’s numerical comparison (PSNR/SSIM) to reconstruct 8 × 8 out of 2 × 2 views: extrapolation 0.
Table 2. The proposed model’s numerical comparison (PSNR/SSIM) to reconstruct 8 × 8 out of 2 × 2 views: extrapolation 0.
DatasetWu [31] Kalantari [15]Shi [16]Yeung [46]Zhang [37]Salem [40]Proposed
30 Scenes35.25/0.92840.11/0.97941.12/0.98541.21/0.98241.98/0.98642.33/0.98542.69/0.986
Reflective35.15/0.94037.35/0.95438.10/0.95838.09/0.95938.71/0.96238.86/0.96239.45/0.967
Occlusions31.77/0.88133.21/0.91134.41/0.92934.50/0.92134.76/0.91834.69/0.92235.41/0.928
Average34.06/0.91636.89/0.94837.88/0.95737.93/0.95438.48/0.95538.62/0.95639.18/0.960
Table 3. The proposed model’s numerical comparison (PSNR/SSIM) to reconstruct 8 × 8 out of 2 × 2 views: extrapolation 1.
Table 3. The proposed model’s numerical comparison (PSNR/SSIM) to reconstruct 8 × 8 out of 2 × 2 views: extrapolation 1.
DatasetYeung [46]Zhang [37]Salem [40]Proposed
30 Scenes42.47/0.98543.57/0.98943.76/0.98844.26/0.989
Reflective41.61/0.97342.33/0.97542.44/0.97443.16/0.979
Occlusions37.28/0.93437.61/0.93737.93/0.94838.47/0.943
Average40.45/0.96441.17/0.96741.38/0.97041.96/0.970
Table 4. The proposed model’s numerical comparison (PSNR/SSIM) to reconstruct 8 × 8 out of 2 × 2 views: extrapolation 2.
Table 4. The proposed model’s numerical comparison (PSNR/SSIM) to reconstruct 8 × 8 out of 2 × 2 views: extrapolation 2.
DatasetYeung [46]Zhang [37]Salem [40]Proposed
30 Scenes42.74/0.98643.41/0.98943.43/0.98743.92/0.989
Reflective41.52/0.97242.09/0.97542.26/0.97542.81/0.978
Occlusions36.96/0.93737.60/0.94437.91/0.94538.25/0.935
Average40.41/0.96541.03/0.96941.20/0.96941.66/0.967
Figure 6. Visual comparison of the proposed model to reconstruct 8 × 8 views out of 2 × 2: extrapolation 0.
Figure 6. Visual comparison of the proposed model to reconstruct 8 × 8 views out of 2 × 2: extrapolation 0.
Sensors 22 05254 g006
Figure 7. Visual comparison of the proposed model to reconstruct 8 × 8 views out of 2 × 2: extrapolation 1 at an interpolated view.
Figure 7. Visual comparison of the proposed model to reconstruct 8 × 8 views out of 2 × 2: extrapolation 1 at an interpolated view.
Sensors 22 05254 g007
Figure 8. Visual comparison of the proposed model to reconstruct 8 × 8 views out of 2 × 2: extrapolation 1 at an extrapolated view.
Figure 8. Visual comparison of the proposed model to reconstruct 8 × 8 views out of 2 × 2: extrapolation 1 at an extrapolated view.
Sensors 22 05254 g008

4.2. Reconstruction Time

Table 5 presents the average run-time to reconstruct a full LF image for the first task: 7 × 7 out of 3 × 3 views. We tested our model on an NVIDIA Geforce RTX 3090. The proposed model can reconstruct LF images faster due to its highly parallel design.
Wang et al. [47] consume a lot of time as they do not reconstruct the entire scene in one feedforward pass. Instead, they reconstruct rows and then columns hierarchically. Yeung et al. [46] and Liu et al. [36] used MATLAB to build their code, which contains many time-consuming reshaping operations. Compared to Salem et al. [38], they used 15 residual blocks (RBs) compared to the 10 RBs in our proposed work. In addition, they process LFs in the raw representation of size 7H × 7W compared to H × W in our implementation.
Table 5. Average run-time to reconstruct 7 × 7 out of 3 × 3 views.
Table 5. Average run-time to reconstruct 7 × 7 out of 3 × 3 views.
Wang [47]Yeung [46]Liu [36] Salem [38]Proposed
Run-Time5.74 s4.58 s2.45 s1.911 s0.686 s

4.3. Ablation Study

We compared three different architectures to validate the effect of the channel attention (CA) mechanism on the reconstruction process. Numerical comparison is presented in Table 6, where the first row indicates the simplest case without applying the CA mechanism. The second row gives the results for the block that is the same as the one proposed in [39] with the CA integrated inside the RCAB. The final row gives the results for the proposed block with the CA separated from the RB, as shown in Figure 9.
Table 6. Investigating the channel attention mechanism (CA) effect on the proposed architecture.
Table 6. Investigating the channel attention mechanism (CA) effect on the proposed architecture.
Model 3 × 3–7 × 730 ScenesReflectiveOcclusions Average
No CA44.86/0.99044.74/0.98140.06/0.95143.22/0.974
CA inside RB46.20/0.99245.71/0.98441.35/0.95444.42/0.976
CA separated from RB46.41/0.99245.73/0.98441.41/0.95144.51/0.976
Figure 9. Different designs of the residual channel attention block (RCAB); (a) simple residual block (RB) without the channel attention mechanism; (b) the proposed RCAB with the CA integrated inside [39]; (c) the proposed RCAB with a CA separated from the RB.
Figure 9. Different designs of the residual channel attention block (RCAB); (a) simple residual block (RB) without the channel attention mechanism; (b) the proposed RCAB with the CA integrated inside [39]; (c) the proposed RCAB with a CA separated from the RB.
Sensors 22 05254 g009

5. Future Work

In this paper, we present a method for reconstructing light field images. The proposed method is characterized by its applicability to all reconstruction tasks for LF images with a small baseline. Although this model is efficient, it fails to reconstruct LF images with a broad baseline. In addition, it sometimes fails to reconstruct parts of the scenes with complex backgrounds or contains severe reflections. Therefore, we are trying to develop a method capable of reconstructing complex scenes and scenes with a broad baseline.

6. Conclusions

This research proposes an effective learning-based paradigm for increasing the angular resolution of LF images. We up-sampled input EPIs to the required angular size, which allows our network to be used for any reconstruction task. In addition, this allowed the network to comprehend and accurately represent the connection since the input and output were of the same size. Finally, we adopted the channel attention mechanism to help the network to concentrate on the important features by assigning higher weights. The proposed RCA_LF network reconstructs LF images by mapping the up-sampled low-resolution images into high-resolution 3D LF volumes. The RCA_LF outperforms other state-of-the-art methods in reconstructing LF images with a small baseline.

Author Contributions

Conceptualization, A.S. and H.I.; methodology, A.S. and H.I.; software, A.S.; formal analysis, A.S.; investigation, H.-S.K.; resources, H.-S.K.; data curation, A.S.; writing—original draft preparation, A.S.; writing—review and editing, A.S. and H.-S.K.; validation, A.S. and H.-S.K.; visualization, H.-S.K.; supervision, H.-S.K.; project administration, H.-S.K.; funding acquisition, H.-S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Basic Science Research Program through the National Research Foundation of Korea (NRF), funded by the Ministry of Education under Grant 2020R1I1A3A04037680 and in part supported by the MSIT (Ministry of Science and ICT), Korea, under the Grand Information Technology Research Center support program (IITP-2022-2020-0-01462) supervised by the IITP (Institute for Information & communications Technology Planning & Evaluation).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this paper are public datasets. We also provide the proposed method’s training and evaluation codes at: https://github.com/ahmeddiefy/RCA-LF, which was created (accessed on 24 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Y.; Wu, T.; Yang, J.; Wang, L.; An, W.; Guo, Y. DeOccNet: Learning to see through foreground occlusions in light fields. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 118–127. [Google Scholar]
  2. Li, Y.; Yang, W.; Xu, Z.; Chen, Z.; Shi, Z.; Zhang, Y.; Huang, L. Mask4D: 4D convolution network for light field occlusion removal. In Proceedings of the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 2480–2484. [Google Scholar]
  3. Wang, W.; Lin, Y.; Zhang, S. Enhanced spinning parallelogram operator combining color constraint and histogram integration for robust light field depth estimation. IEEE Signal Process. Lett. 2021, 28, 1080–1084. [Google Scholar] [CrossRef]
  4. Wang, Y.; Wang, L.; Liang, Z.; Yang, J.; An, W.; Guo, Y. Occlusion-Aware Cost Constructor for Light Field Depth Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 19809–19818. [Google Scholar]
  5. Shi, J.; Jiang, X.; Guillemot, C. A framework for learning depth from a flexible subset of dense and sparse light field views. IEEE Trans. Image Process. 2019, 28, 5867–5880. [Google Scholar] [CrossRef] [Green Version]
  6. Zhang, J.; Liu, Y.; Zhang, S.; Poppe, R.; Wang, M. Light field saliency detection with deep convolutional networks. IEEE Trans. Image Process. 2020, 29, 4421–4434. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Piao, Y.; Jiang, Y.; Zhang, M.; Wang, J.; Lu, H. PANet: Patch-Aware Network for Light Field Salient Object Detection. IEEE Trans. Cybern. 2021. [Google Scholar] [CrossRef] [PubMed]
  8. Lin, G.; Zhu, L.; Li, J.; Zou, X.; Tang, Y. Collision-free path planning for a guava-harvesting robot based on recurrent deep reinforcement learning. Comput. Electron. Agric. 2021, 188, 106350. [Google Scholar] [CrossRef]
  9. Lin, G.; Tang, Y.; Zou, X.; Wang, C. Three-dimensional reconstruction of guava fruits and branches using instance segmentation and geometry analysis. Comput. Electron. Agric. 2021, 184, 106107. [Google Scholar] [CrossRef]
  10. Wu, G.; Masia, B.; Jarabo, A.; Zhang, Y.; Wang, L.; Dai, Q.; Chai, T.; Liu, Y. Light field image processing: An overview. IEEE J. Sel. Top. Signal Process. 2017, 11, 926–954. [Google Scholar] [CrossRef] [Green Version]
  11. Georgiev, T.G.; Lumsdaine, A. Focused plenoptic camera and rendering. J. Electron. Imaging 2010, 19, 021106. [Google Scholar]
  12. Raytrix. Available online: https://raytrix.de/ (accessed on 6 May 2022).
  13. Wu, G.; Zhao, M.; Wang, L.; Dai, Q.; Chai, T.; Liu, Y. Light field reconstruction using deep convolutional network on EPI. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6319–6327. [Google Scholar]
  14. Hu, Z.; Chung, Y.Y.; Ouyang, W.; Chen, X.; Chen, Z. Light field reconstruction using hierarchical features fusion. Expert Syst. Appl. 2020, 151, 113394. [Google Scholar] [CrossRef]
  15. Kalantari, N.K.; Wang, T.-C.; Ramamoorthi, R. Learning-based view synthesis for light field cameras. ACM Trans. Graph. (TOG) 2016, 35, 1–10. [Google Scholar] [CrossRef] [Green Version]
  16. Shi, J.; Jiang, X.; Guillemot, C. Learning fused pixel and feature-based view reconstructions for light fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–20 June 2020; pp. 2555–2564. [Google Scholar]
  17. Cheng, Z.; Xiong, Z.; Chen, C.; Liu, D.; Zha, Z.-J. Light field super-resolution with zero-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 10010–10019. [Google Scholar]
  18. Ko, K.; Koh, Y.J.; Chang, S.; Kim, C.-S. Light field super-resolution via adaptive feature remixing. IEEE Trans. Image Process. 2021, 30, 4114–4128. [Google Scholar] [CrossRef] [PubMed]
  19. Zhang, S.; Chang, S.; Lin, Y. End-to-end light field spatial super-resolution network using multiple epipolar geometry. IEEE Trans. Image Process. 2021, 30, 5956–5968. [Google Scholar] [CrossRef] [PubMed]
  20. Bergen, J.R.; Adelson, E.H. The plenoptic function and the elements of early vision. Comput. Models Vis. Process. 1991, 1, 8. [Google Scholar]
  21. Levoy, M.; Hanrahan, P. Light field rendering. In Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, LA, USA, 4–9 August 1996; pp. 31–42. [Google Scholar]
  22. Wanner, S.; Goldluecke, B. Spatial and angular variational super-resolution of 4D light fields. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 608–621. [Google Scholar]
  23. Mitra, K.; Veeraraghavan, A. Light field denoising, light field superresolution and stereo camera based refocussing using a GMM light field patch prior. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 18–20 June 2012; pp. 22–28. [Google Scholar]
  24. Pujades, S.; Devernay, F.; Goldluecke, B. Bayesian view synthesis and image-based rendering principles. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 3906–3913. [Google Scholar]
  25. Chaurasia, G.; Duchene, S.; Sorkine-Hornung, O.; Drettakis, G. Depth synthesis and local warps for plausible image-based navigation. ACM Trans. Graph. 2013, 32, 1–12. [Google Scholar] [CrossRef]
  26. Zhang, F.-L.; Wang, J.; Shechtman, E.; Zhou, Z.-Y.; Shi, J.-X.; Hu, S.-M. Plenopatch: Patch-based plenoptic image manipulation. IEEE Trans. Vis. Comput. Graph. 2016, 23, 1561–1573. [Google Scholar] [CrossRef] [Green Version]
  27. Vagharshakyan, S.; Bregovic, R.; Gotchev, A. Light field reconstruction using shearlet transform. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 133–147. [Google Scholar] [CrossRef] [Green Version]
  28. Salem, A.; Ibrahem, H.; Kang, H.-S. Dual Disparity-Based Novel View Reconstruction for Light Field Images Using Discrete Cosine Transform Filter. IEEE Access 2020, 8, 72287–72297. [Google Scholar] [CrossRef]
  29. Jin, J.; Hou, J.; Yuan, H.; Kwong, S. Learning light field angular super-resolution via a geometry-aware network. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 11141–11148. [Google Scholar]
  30. Yeung, H.W.F.; Hou, J.; Chen, X.; Chen, J.; Chen, Z.; Chung, Y.Y. Light field spatial super-resolution using deep efficient spatial-angular separable convolution. IEEE Trans. Image Process. 2018, 28, 2319–2330. [Google Scholar] [CrossRef]
  31. Wu, G.; Liu, Y.; Fang, L.; Dai, Q.; Chai, T. Light field reconstruction using convolutional network on EPI and extended applications. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1681–1694. [Google Scholar] [CrossRef]
  32. Wu, G.; Liu, Y.; Fang, L.; Chai, T. Revisiting light field rendering with deep anti-aliasing neural network. IEEE Trans. Pattern Anal. Mach. Intell. 2021. [Google Scholar] [CrossRef]
  33. Meng, N.; So, H.K.-H.; Sun, X.; Lam, E.Y. High-dimensional dense residual convolutional neural network for light field reconstruction. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 873–886. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Mildenhall, B.; Srinivasan, P.P.; Ortiz-Cayon, R.; Kalantari, N.K.; Ramamoorthi, R.; Ng, R.; Kar, A. Local light field fusion: Practical view synthesis with prescriptive sampling guidelines. ACM Trans. Graph. 2019, 38, 1–14. [Google Scholar] [CrossRef]
  35. Wang, Y.; Liu, F.; Zhang, K.; Wang, Z.; Sun, Z.; Tan, T. High-fidelity view synthesis for light field imaging with extended pseudo 4DCNN. IEEE Trans. Comput. Imaging 2020, 6, 830–842. [Google Scholar] [CrossRef]
  36. Liu, D.; Huang, Y.; Wu, Q.; Ma, R.; An, P. Multi-angular epipolar geometry based light field angular reconstruction network. IEEE Trans. Comput. Imaging 2020, 6, 1507–1522. [Google Scholar] [CrossRef]
  37. Zhang, S.; Chang, S.; Shen, Z.; Lin, Y. Micro-Lens Image Stack Upsampling for Densely-Sampled Light Field Reconstruction. IEEE Trans. Comput. Imaging 2021, 7, 799–811. [Google Scholar] [CrossRef]
  38. Salem, A.; Ibrahem, H.; Kang, H.-S. Light Field Reconstruction Using Residual Networks on Raw Images. Sensors 2022, 22, 1956. [Google Scholar] [CrossRef]
  39. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  40. Salem, A.; Ibrahem, H.; Yagoub, B.; Kang, H.-S. End-to-End Residual Network for Light Field Reconstruction on Raw Images and View Image Stacks. Sensors 2022, 22, 3540. [Google Scholar] [CrossRef]
  41. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  42. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  43. Raj, A.S.; Lowney, M.; Shah, R.; Wetzstein, G. Stanford Lytro Light Field Archive; Stanford Computational Imaging Lab: Stanford, CA, USA, 2016. [Google Scholar]
  44. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M. {TensorFlow}: A System for {Large-Scale} Machine Learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  45. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  46. Yeung, H.W.F.; Hou, J.; Chen, J.; Chung, Y.Y.; Chen, X. Fast light field reconstruction with deep coarse-to-fine modeling of spatial-angular clues. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 137–152. [Google Scholar]
  47. Wang, Y.; Liu, F.; Wang, Z.; Hou, G.; Sun, Z.; Tan, T. End-to-end view synthesis for light field imaging with pseudo 4DCNN. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 333–348. [Google Scholar]
Figure 1. (a) The 4D LF representation with two planes’ intersections of light rays. (b) LF images.
Figure 1. (a) The 4D LF representation with two planes’ intersections of light rays. (b) LF images.
Sensors 22 05254 g001
Figure 2. (a) The 4D LF representation as a 2D array of view images. (b) The 3D LF view image stack.
Figure 2. (a) The 4D LF representation as a 2D array of view images. (b) The 3D LF view image stack.
Sensors 22 05254 g002
Figure 3. (a) Overview of the proposed RCA-LF network structure. (b) Implementation details of the residual channel attention block (RCAB).
Figure 3. (a) Overview of the proposed RCA-LF network structure. (b) Implementation details of the residual channel attention block (RCAB).
Sensors 22 05254 g003
Figure 4. The input–output relationship for different LF reconstruction tasks. (a) 3 × 3–7 × 7. (b) 2 × 2–8 × 8, extra 0. (c) 2 × 2–8 × 8, extra 1. (d) 2 × 2–8 × 8, extra 2.
Figure 4. The input–output relationship for different LF reconstruction tasks. (a) 3 × 3–7 × 7. (b) 2 × 2–8 × 8, extra 0. (c) 2 × 2–8 × 8, extra 1. (d) 2 × 2–8 × 8, extra 2.
Sensors 22 05254 g004
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Salem, A.; Ibrahem, H.; Kang, H.-S. RCA-LF: Dense Light Field Reconstruction Using Residual Channel Attention Networks. Sensors 2022, 22, 5254. https://doi.org/10.3390/s22145254

AMA Style

Salem A, Ibrahem H, Kang H-S. RCA-LF: Dense Light Field Reconstruction Using Residual Channel Attention Networks. Sensors. 2022; 22(14):5254. https://doi.org/10.3390/s22145254

Chicago/Turabian Style

Salem, Ahmed, Hatem Ibrahem, and Hyun-Soo Kang. 2022. "RCA-LF: Dense Light Field Reconstruction Using Residual Channel Attention Networks" Sensors 22, no. 14: 5254. https://doi.org/10.3390/s22145254

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop