Next Article in Journal
Electrochemical Analysis of Attoliter Water Droplets in Organic Solutions through Partitioning Equilibrium
Previous Article in Journal
Multi-Robot Exploration of Unknown Space Using Combined Meta-Heuristic Salp Swarm Algorithm and Deterministic Coordinated Multi-Robot Exploration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Event-Guided Image Super-Resolution Reconstruction

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
College of Materials Science and Opto-Electronic Technology, University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 2155; https://doi.org/10.3390/s23042155
Submission received: 10 January 2023 / Revised: 6 February 2023 / Accepted: 10 February 2023 / Published: 14 February 2023
(This article belongs to the Topic 3D Computer Vision and Smart Building and City)

Abstract

:
The event camera efficiently detects scene radiance changes and produces an asynchronous event stream with low latency, high dynamic range (HDR), high temporal resolution, and low power consumption. However, the large output data caused by the asynchronous imaging mechanism makes the increase in spatial resolution of the event camera limited. In this paper, we propose a novel event camera super-resolution (SR) network (EFSR-Net) based on a deep learning approach to address the problems of low spatial resolution and poor visualization of event cameras. The network model is capable of reconstructing high-resolution (HR) intensity images using event streams and active sensor pixel (APS) frame information. We design the coupled response blocks (CRB) in the network that are able of fusing the feature information of both data to achieve the recovery of detailed textures in the shadows of real images. We demonstrate that our method is able to reconstruct high-resolution intensity images with more details and less blurring in synthetic and real datasets, respectively. The proposed EFSR-Net can improve the peak signal-to-noise ratio (PSNR) metric by 1–2 dB compared with state-of-the-art methods.

1. Introduction

The event camera is a new bio-inspired vision sensor with the pixel structure shown in Figure 1 [1]. The pixel structure is mainly divided into three parts, including a voltage–current logarithmic conversion unit, a change amplification unit and a comparison unit. Event cameras work completely differently than traditional cameras. They do not capture images at a constant rate, but only output information about changes in local pixel brightness. The event generation process is shown in Figure 2 [2]. When the light intensity changes beyond a set threshold, the event camera marks the timestamp with microsecond time resolution and outputs the event. The event information will encode the time, position, and polarity of this change, and the event can be simply represented as a tuple e ( x , y , t , p ) , where x, y are the coordinates of the output pixel points, t is the timestamp of the event, and p is the polarity representing the luminance change (polarity information can be defined as a function shown in Equation (1) [3]).
p = { 1     1 i f   O F F   e v e n t i f   o n   e v e n t
A series of events eventually form the event stream, which is shown in Equation (2).
E = i = 1 N e ( x i , y i , t i , p i )
Due to its unique way of working, the event camera has advantages that traditional cameras do not have, such as low latency, high dynamic range (HDR), high temporal resolution, and low power consumption. Therefore, the event camera is suitable for extreme situations such as high-speed motion and large changes in lighting conditions, making it a research hotspot in robotics and computer vision [4,5,6,7,8].
Considering the need for visualization, the output of the latest dynamic and active pixel vision sensor (DAVIS [6]) combines asynchronous event streams and synchronous image frames. However, on the one hand, the spatial resolution of event cameras is generally not high due to the consideration of data transmission efficiency, i.e., there is a balance between spatial resolution and latency. To meet the low latency data characteristics of event cameras, event cameras are able to capture the luminance changes of each pixel individually [6]. This also leads to its increased noise sensitivity, making the visualization less effective. In summary, reconstructing high-quality intensity images from the output of an event camera and thus meeting the visualization needs of the camera is a significant issue. This will facilitate the application of event cameras to many advanced vision tasks [9,10,11,12], such as recognition, detection, and tracking, that are solved by standard cameras.
Most traditional super-resolution reconstruction methods directly reconstruct intensity images from event stream data [13,14,15,16]. However, the quality of reconstructed images is limited by the sparse nature and low resolution of event stream data, and the intensity information is ignored. Some super-resolution reconstruction methods start to use event data as an aid to super-resolve intensity images. However, due to degradation problems such as blurring and noise for various reasons, they may not produce high-quality images and cannot recover more details in the shadows of real images by using the high dynamic range properties of event data.
In this paper, we propose the EFSR-Net neural network model. It can reconstruct high-quality intensity images with more recovered details and higher dynamic range by using both event data and APS frames as data input. Overall, the main contributions of this paper are as follows:
  • We designed a novel network model suitable for super-resolution reconstruction of intensity images from event cameras, named EFSR-Net. Our algorithm is based on a hybrid paradigm of frames and events. The final super-resolution effect is significantly better than that of simply reconstructing from a low-resolution event stream as input;
  • We designed the coupled response block (CRB) in the network. It can fuse the event data and APS frame data to complement each other, and recover the texture details contained in the real image shadows by using the high dynamic range characteristics of the event data.
The rest of this paper is organized as follows. First, we focus on developing traditional methods for super-resolution reconstruction of event camera images and their respective advantages and disadvantages in Section 2. Secondly, the proposed method is described in Section 3, and the details of our proposed network are systematically presented and analyzed. Next, in Section 4, our experimental part is presented, comparing our method with other state-of-the-art methods and giving qualitative and quantitative analyses to verify the effectiveness of our approach. Finally, conclusions are given in Section 5.

2. Related Work

2.1. Event Data Processing Method

The event camera converts motion changes into a spatially sparse, temporally dense stream of events. These data contain dynamic information about the edges of the object. However, the event streams are represented in terms of logarithmic intensity changes, which have a completely different data format than the general intensity images [17]. Therefore, the event data are generally processed when reconstruction operations are performed. There have been many approaches to event stream representation and processing, which can be broadly classified into two types based on manual production [15,16,18], and end-to-end [19,20,21]. The primary method of representing events based on manual display uses frequency accumulation images. Frequency accumulation images are divided into fixed-time accumulation images and fixed-number-of-events accumulation images. In the specified time accumulation, the event streams between two consecutive intensity images (APS) of the event camera corresponding to the reference time is divided into n equally scaled parts, forming n frames. These n frames are stacked to eventually form a stack that is fed to the network as input. This approach preserves the temporal dimension information of the event data to a greater extent. However, this stacking method has a problem of missing events when the scene or camera is not moving. When there is not enough event data in the time interval for image reconstruction, it is inevitably difficult to obtain good super-resolution reconstructed images. Also, the number of events in different time intervals may appear excessive. The fixed number of accumulated images is an excellent way to avoid the above problem. It forms frames by merging events according to the number of incoming events. The first N e events are merged into frame 1, the next N e events are merged into frame 2, and so on, creating n frames that form a stack. Since we count the number of events over time, we can adaptively adjust the number of events in each frame as well as in a stack.
End-to-end event representation methods mainly include grid-based representation, which samples event streams into a spatial–temporal voxel grid, and point-set-based representation, which is treated as a point cloud. They can both use end-to-end neural networks, and supervised learning driven by big data can better mine the spatial–temporal properties of event stream signals. However, these types of methods cannot manipulate the data and cannot select well near the timestamp we are interested in.

2.2. Event-Based Intensity Reconstruction

Due to its wide range of applications, the reconstruction of intensity images from events has become a hot research topic in event cameras. An earlier attempt to study intensity reconstruction from pure events was to consider relatively short-period event stream data and directly accumulate positive and negative events of two colors as the output of the gradient interpretation. Hanme Kim et al. proposed to replace the gradient representation with a synthetic intensity image using simultaneous estimation of camera motion and stitching it into a panoramic gradient image. In this method, the scene is static, with only rotational motion of the camera. The gradient image is finally converted to an intensity image by Poisson integration [20]. The reconstruction of HR images based on spherical 3D scenes was further investigated in depth in [12]. In [21], Cook et al. proposed a cyclic structured bionic interconnection network that achieves simultaneous reconstruction of intensity frames, optical flow, and angular velocity of small rotational motions. Bardow et al. [22] proposed the joint estimation of intensity variations and optical flow in a unified variational energy minimum framework in a challenging dynamic motion environment. The optimization allows simultaneous reconstruction of video frames and optical streams, but this method propagates errors as shadow-like artifacts in the generated intensity images. A variational framework based on a denoising scheme that iteratively filters incoming events was introduced in [23]. The method uses a popular regularization on the relative timestamps of the events to recover more details in texture-free regions to achieve reconstructed images. In [14], a high-pass filter was proposed to reconstruct the video in a very efficient way. The framework was originally designed to complement intensity frames with event information, but can also reconstruct images from events without the help of APS frames. In contrast to [14], Munda et al. [23] considered image reconstruction as defined in the energy minimization problem induced by the event timestamp. In recent years, deep learning-based methods have greatly progressed in intensity image and video reconstruction. Wang et al. [15] took to reconstructing intensity images using generative adversarial networks on the U-Net [24] architecture. Rebecq et al. [13] used recurrent neural networks and stacked ConvLSTM gates to reconstruct videos from events. The above methods all perform intensity reconstruction from event stream data, while APS frames contain relatively rich textures, so events and APS frames can be used as complementary sources for event-based intensity reconstruction. In [14], events are approximated as temporal differences in intensity frames. On this basis, a complementary filter is proposed as a fusion engine that can generate intensity frames almost continuously. Pan et al. [25] proposed an event-based deblurring method by correlating blurred APS frames and events with an event-based double integration (EDI) model. Next, a multi-frame EDI model for high-speed video reconstruction is proposed by further considering the relationship between frames [26]. Zhang et al. [27] formulated event-based image reconstruction as a linear inverse problem with deep regularization using optical flow. They have emphasized the framework of simultaneously estimating both physically entangled quantities in the events: brightness and motion (optical flow).

2.3. Event-Based Super-Resolution

Since the spatial resolution of event cameras is limited, some work in recent years has focused on the super-resolution of event streams in the spatial and temporal domains. Mohammad et al. [16] were the first to design recurrent neural networks to solve the super-resolution event camera problem. They created an SRNet network to perform super-resolution reconstruction from pure event data. Li et al. [28] used Poisson point processes to model event sequences and sampled the events according to a nonhomogeneous Poisson process. Duan et al. [29] proposed EventZoom, which is a deep neural network framework with a three-dimensional U-Net [24] backbone architecture that addresses joint denoising and super-resolution of neuromorphic events. Wang et al. [18] proposed an end-to-end network called EventSR, which implements image reconstruction, enhancement, and upsampling from the event streams. The network is trained in three stages and uses unsupervised adversarial learning. They also created an open dataset. Wang et al. [30] designed a sparse learning network for event enhancement to simultaneously solve the deblurring, denoising and super-resolution problems. In [31], a hybrid camera was constructed to guide the event filtering and take advantage of the HR RGB signal to guide the upsampling of events. In [17], Han et al. converted event data into potential intensity frames and reconstructed SR intensity images and high frame rate videos with higher dynamic range and less blurring artifacts by fusing potential frames with APS frames. Weng et al. [32] proposed a recurrent neural network for event SR without the assistance of frames. They also demonstrated the feasibility of achieving event SR solely with pure events. Song et al. [33] proposed E-CIR, which converts a blurry image into a sharp video represented as a parametric function from time to intensity. It achieves good results in deblurring but still lacks in improving image resolution.

3. Proposed Method

3.1. Overview

In this paper, we design a convolutional neural network EFSR-Net consisting of several modules. Figure 3 shows the overall structure of our proposed EFSR-Net. Our network takes mixed types of data as input and fuses intensity frame information with event stream information to achieve super-resolution reconstruction of high-quality intensity images from event cameras. Our approach is divided into two main steps. The event stream is first preprocessed with data, and we choose event stacking for processing, as in [16]. This method is able to select event data near the timestamp we are interested in. In the second step, the APS frames and the stack containing the event information are passed through the neural network as input for super-resolution reconstruction. Our super-resolution network consists of upper and lower coupled sub-networks for feature extraction on event frames and APS frames, respectively. In particular, we designed a coupled reaction block (CRB) to achieve the fusion and complementarity of the two data features. In Section 3.2, we describe the preprocessing method for event data. In Section 3.3, the network structure is introduced, explaining the details of the whole network architecture. In Section 3.4, the loss function of our network is described.

3.2. Event Data Preprocessing

The event streams data has HDR properties and high temporal resolution. We aim to perform intensity image super-resolution reconstruction using APS frames combined with event data around the corresponding timestamps. Event streams represent sparse logarithmic intensity changes and have a completely different type of data format than ordinary intensity images. Therefore, it is not easy to fuse event data and intensity image information directly. In [34], it is stated that under ideal conditions (noise-free scenes, perfect sensor response, etc.), the integration of events produces “absolute” luminance, i.e., it is possible to integrate intensity images from event information, since events simply encode the visual content of the scene in a non-redundant way for each pixel. In addition, due to the high temporal resolution of the events, the luminance images can meet a very high frame rate for reconstruction [14].
Since we need to obtain the event stream data around the corresponding APS frame, as mentioned earlier, end-to-end event processing methods based on grids, etc., are not suitable. Therefore, this paper uses a stacking approach based on the number of events, where a fixed number of events are stacked to form a stack. The specific stacking method is to start counting the number of events according to the timestamp information until the predefined number of events N e is reached, and the accumulated events form a channel in the stack. The above work is repeated C times to form C channels. We make the size of the stack consistent with the size of the APS frame and finally obtain some column stacks S i H × W × C ,   i = 1 ,   2 ,   3 , . Therefore, each stack contains M = C × N e events. In this paper, we set C = 3, N e   = 5000, H = 320 and W = 240.

3.3. Network Architecture

Inspired by [15], we used the basic architecture of U-Net [24] to finally include our event data in the intensity image through an encoder and decoder to facilitate the processing of event data by our subsequent network. We performed a series of encoding and decoding as well as hopping connections on the stack S H × W × C formed by stacking events, and finally output the intermediate intensity image I e i H × W × 1 . Next, we fed the intermediate intensity image data containing the event data and the APS frame data to our super-resolution network. Our super-resolution network consists of two coupled sub-networks, upper and lower, taking the processed event data and the APS frame image data as input, respectively. Each sub-network contains an initial feature extraction block (FEB), a super-resolution block (SRB), a coupled response block (CRB), and a reconstruction block (REB). The APS frame data is denoted as I a i . The corresponding features extracted by FEB can be expressed as:
L e = f F E B ( I e i )
L a = f F E B ( I a i )
where f F E B indicates the operation of the feature extraction block.
The FEB uses a convolutional layer with PReLU activation, consisting of 256 filters sized 3 × 3 precisely, to extract the features of the image. The extracted features L e , L a are used as the basic inputs for the subsequent SRB and CRB. Using the basic features L e , L a as input, the role of the SRB is to learn more advanced features and improve the image resolution. The feedback architecture of the SRB follows [35]. The advanced features learned by the SRB can be expressed as:
F e = f S R B ( L e )
F a = f S R B ( L a )
where f S R B denotes the operation of the SRB. F e and F a denote the intermediate intensity image containing event information and the advanced features extracted from the APS frame image, respectively, which are important inputs for the subsequent CRB.
The coupled response block (CRB) enables simultaneous super-resolution and fusion of information from two data sources through complex network connections. We take the outputs F e and F a of the SRB module and the features L e and L a extracted at the FEB as inputs, first connected by a set of 1 × 1 filters, and then perform a series of upsampling and downsampling operations to extract the feature map. The output F e , F a of the SRB module is used to correct the basic characteristics of L e , L a and improve the performance of SR. Furthermore, F e , F a bring complementary information to each sub-network to achieve feature fusion. The output can be expressed as:
M e = f C R B ( F e , F a , L e )
M a = f C R B ( F a , F e , L a )
where f C R B denotes the operation of the CRB. M e , M a are the outputs of the CRB blocks in the upper and lower sub-networks.
After the data stream passes through the upper and lower two CRBs, the information is reconstructed in the next step. The reconstruction can be expressed as follows:
I e o = f U P ( I e i ) + f R E B ( M e )
I a o = f U P ( I a i ) + f R E B ( M a )
here I e o and I a o are the intermediate output intensity information of the upper and lower subnetworks, respectively. f U P denotes the upsampling operation and f R E B denotes the reconstruction block operation.
Finally, the reconstructed features are fused by our mixer (Mix) using convolutional layers to finally generate super-resolution intensity images. The output can be expressed as follows:
I o u t = w e I e + w a I a
where I o u t is the reconstructed image we finally obtain, and the initial values of w e and w a are chosen as 0.5.

3.4. Loss Function

There are many choices of loss functions used for optimization, such as L1 loss function, L2 loss function, perceptual loss function, perceptual similarity loss function, and so on. Because the L2 loss function causes the output image to become smooth, we choose the L1 loss function in order to get a clearer image. However, the L1 loss can only characterize the low-level feature differences of the image, which can easily cause the image not to match the visual perception of the human eye. So we add perceptual similarity loss [36] to the L1 loss function as in [16]. We extract the feature stack from layer l and perform unit normalization in the channel dimension, and the perceptual similarity loss is calculated as follows:
L p s = l 1 H l W l h , w w l ( y ^ h w l y ^ 0 h w l ) 2 2
where y ^ l , y ^ 0 l H l × W l × C l denotes the l-layer channel dimension. The vector w l C l . The total loss function is:
L l o s s = L 1 + λ L p s
where λ is the balance parameter. The network is trained by minimizing the loss function to bring the output closer to the desired image.

4. Experiment

4.1. Dataset Preparation

We need APS frame and event stream data and the corresponding ground truth image sequences for network training. However, the dataset collected by the real event sensor can only provide low-resolution images of poor quality, which cannot be used as the ground truth images needed for training. Therefore, we trained the network on a synthetic dataset we borrowed from [30] and processed it. It selects clear images with a resolution of 1280 × 720 from the Gopro dataset [37] as the ground truth (GT) images and generates events from a series of input images using the event camera simulator (ESIM) [38]. We choose 120 video sequences, each containing 95 images, for a total of 11,400 images. We generate the corresponding low-resolution images in our experiments by sampling HR images with bicubic interpolation. Their resolutions are 640 × 360 and 320 × 180, corresponding to our 4× and 2× training, respectively. In the experiments, 80% of the dataset is used for training and 20% for testing.

4.2. Implementation Details

During the training process, the initial learning rate is set to 0.01, the total number of training units is 80, and the balance parameter λ is 0.3. After 50 epochs, the learning rate decays to 0.001, and the batch size is set to 8. The optimizer uses the Adam [39] algorithm with parameters β 1 = 0.9 and β 2 = 0.999.
We implemented our proposed model in Pytorch version 1.12.0 and Python version 3.9, which were trained using a single NVIDIA GeForce RTX 3090 GPU. the CUDA and CuDNN versions are 11.6 and 8.4.0, respectively.

4.3. Compare with Advanced Algorithms

In this work, we compare the super-resolution algorithm with state-of-the-art event cameras on synthetic and real datasets, respectively, to demonstrate the superiority of our approach and its generalizability to real scenarios. To my knowledge, there are only a few super-resolution algorithms for event cameras, so, like other related articles [16,17], we also compare the reconstruction algorithm for event cameras combined with the image-based super-resolution algorithm. The comparison is as follows:
  • EV [15] + SISR [40], the method first reconstructs the intensity image by E2VID [15] using the event streams data as input and then super-resolves it with the trained single image super-resolution (SISR) network [40];
  • E2SRI [16], the method uses pure event data as input to super-resolve the reconstructed intensity image;
  • eSL-Net [30], which uses APS frames and event data to reconstruct HR intensity images super-resolved.
We demonstrate the superiority of our proposed method from both qualitative and quantitative aspects, respectively. For the quantitative analysis of the experiments, we choose peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) as our evaluation metrics. The following is a detailed description of the evaluation metrics:
The peak signal-to-noise ratio (PSNR) can be expressed as:
P S N R ( Y , Y ^ ) = 10 log 10 ( max ( Y ) 2 M S E ( Y , Y ^ ) )
where Y ^ represents the generated image and Y represents the original image, MSE is the mean squared error; the larger the PSNR, the better the quality of the image.
The structural similarity (SSIM) is a measure of how similar two images are, and can be expressed as:
S S I M ( Y , Y ^ ) = ( 2 μ Y μ Y ^ + c 1 ) ( 2 σ Y Y ^ + c 2 ) ( μ Y 2 + μ Y ^ 2 + c 1 ) ( σ Y 2 + σ Y ^ 2 + c 2 )
where μ Y is the mean of Y, μ Y ^ is the mean of Y ^ , σ Y 2 is the variance of Y, σ Y ^ 2 is the variance of Y ^ , and σ Y Y ^ is the covariance of Y and Y ^ . c 1 = ( k 1 L ) 2 , c 1 = ( k 2 L ) 2 are constants used to maintain stability. L is the dynamic range of pixel values. k 1 = 0.01 , k 2 = 0.03 .

4.3.1. Evaluation on Synthetic Datasets

In this section, we use both qualitative and quantitative methods to illustrate the effectiveness of our proposed method in terms of both visual perspective and objective evaluation metrics. Our test set consists of 2280 intensity images and a stream of events between two consecutive frames from 24 high-frame rate video sequences. The quantitative results (PSNR, SSIM) are shown in Table 1, and our method is highly competitive with other popular methods. It should be noted that in performing the comparison, the E2SRI [16] method does not have a 4× SR pre-trained model, so the PSNR and SSIM values are not available in the corresponding positions in Table 1. The eSL-Net [30] method does not have a pre-trained model of 2× SR, and we use bicubic interpolation to process its 4× SR results to obtain the quantitative results corresponding to 2× SR. Compared to the above methods, it can be seen from the quantitative evaluation that our approach is able to reconstruct higher quality 2× and 4× high-resolution intensity images in synthetic dataset experiments. Furthermore, the fusion of intensity images and event data can achieve higher quality image super-resolution with more structural details than reconstructing images from event data alone.
Figure 4 and Figure 5 visualize our comparison results with other methods at 2× SR and 4× SR, respectively, from a visual perspective. A comparison of the reconstruction quality of the different methods in the two scenes from the synthetic dataset shows that the event data contains more edge detail information in the scene. A comparison of the APS frames with the event stacked images in Figure 4 shows that the event data capture more information about the scene and is unaffected by blurring. Comparing the EV [15] + SISR [40] method in Figure 4 with our method, it can be seen that our method is able to reconstruct high-quality images with more continuous grayscale. This indicates that the APS frames provide more continuous grayscale information for our image reconstruction, which complements the drawback of too little event data in scenes with little change in light intensity and thus distortion of the reconstructed image.

4.3.2. Evaluation on Real Dataset

To further illustrate the effectiveness of our proposed algorithm, we compared our method with other methods on seven sequences from the ESIM real dataset [41] as well as sequences from [14]. These sequences were recorded using a DAVIS240C sensor [6] moving in various environments. It contains events as well as grayscale frames at a rate of 20 Hz. Since there are no corresponding high-quality images in the real dataset that can be used as ground truth, we performed qualitative analysis on three sequences of boxes_6dof, office_zigzag, and motorbike from different datasets, and the results are shown in Figure 6, Figure 7 and Figure 8.
As with the comparison method for the synthetic dataset, we compare the 2× SR and 4× SR results of the different methods on sequences of the real dataset separately, and the SR results demonstrate that our method is able to reconstruct more details and sharper edges. Compared to the results on the synthetic dataset, in the real dataset, our method results in images with a higher dynamic range and is able to recover the texture details contained in the shadows of the real images. This is made possible by the high dynamic range nature of the event camera, which allows the event data to contain more information about the scene. From the visual viewpoint, it can be seen that our method has the best visual effect in the real dataset as well. This shows that our method is equally applicable in real scenarios.

4.4. Ablation Experiment

To demonstrate the effectiveness of the perceptual loss function added to our loss function, we conducted ablation experiments by removing the perceptual loss function from the total loss function. From Table 2, we can see that we can improve our evaluation metrics by introducing the perceptual loss function in the loss function, which proves its effectiveness. We choose the slider_depth sequence of the ESIM real dataset [36] to conduct ablation experiments on the variable N e that controls the number of events. This further proves the effectiveness of our method. Our aim is to reconstruct the intensity image with higher resolution and recover more details, so we show the effect of changing N e on the reconstruction quality from a visual perspective, and the results are shown in Figure 9. It can be seen from the figure that a small number of events in the stack will make the event flow information too small to provide more details of the scene, resulting in low reconstruction quality; too many events in the stack will make subsequent events overwrite previous events, resulting in blurring and smearing, resulting in low reconstruction quality.

5. Conclusions

In this paper, we propose a novel network called EFSR-Net for reconstructing high-resolution intensity images from event cameras, which addresses the problem of the low spatial resolution of event cameras. In particular, we propose that the coupled response block (CRB) is able to fuse the event streams with the feature information of APS frames to complement each other. We exploit the high dynamic range properties of the event stream data to enable the reconstructed images to recover the texture details contained in the shadows of the real images. Our experiments on synthetic and real datasets demonstrate the superiority of EFSR-Net, and that EFSR-Net outperforms existing methods in terms of qualitative and quantitative results.

Author Contributions

All authors were involved in the formulation of the problem and the design of the methodology; G.G. designed the experiment and wrote the manuscript; Y.F. constructed the datasets; H.L. (Hailong Liu) and G.B. analyzed the accuracy of the experimental data; Y.Z. and H.L. (Hengyi Lv) reviewed and guided the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (62005269).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Posch, C.; Serrano-Gotarredona, T.; Linares-Barranco, B.; Delbruck, T. Retinomorphic Event-Based Vision Sensors: Bioinspired Cameras with Spiking Output. Proc. IEEE 2014, 102, 1470–1484. [Google Scholar] [CrossRef]
  2. Lichtsteiner, P.; Posch, C.; Delbruck, T. A 128 × 128 120 dB 15 μs Latency Asynchronous Temporal Contrast Vision Sensor. IEEE J. Solid-State Circuits 2008, 43, 566–576. [Google Scholar] [CrossRef]
  3. Reverter Valeiras, D.; Lagorce, X.; Clady, X.; Bartolozzi, C.; Ieng, S.H.; Benosman, R. An Asynchronous Neuromorphic Event-Driven Visual Part-Based Shape Tracking. IEEE Trans. Neural Netw. Learn. Syst. 2015, 26, 3045–3059. [Google Scholar] [CrossRef] [PubMed]
  4. Wang, Y.; Yang, J.; Peng, X.; Wu, P.; Gao, L.; Huang, K.; Chen, J.; Kneip, L.J.S. Visual odometry with an event camera using continuous ray warping and volumetric contrast maximization. Sensors 2022, 22, 5687. [Google Scholar] [CrossRef] [PubMed]
  5. Delbruck, T.; Li, C.; Graca, R.; Mcreynolds, B. Utility and feasibility of a center surround event camera. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 381–385. [Google Scholar]
  6. Brandli, C.; Berner, R.; Minhao, Y.; Shih-Chii, L.; Delbruck, T. A 240 × 180 130 dB 3 µs Latency Global Shutter Spatiotemporal Vision Sensor. IEEE J. Solid-State Circuits 2014, 49, 2333–2341. [Google Scholar] [CrossRef]
  7. Hidalgo-Carrió, J.; Gallego, G.; Scaramuzza, D. Event-aided direct sparse odometry. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–22 June 2022; pp. 5781–5790. [Google Scholar]
  8. Ozawa, T.; Sekikawa, Y.; Saito, H.J.S. Accuracy and Speed Improvement of Event Camera Motion Estimation Using a Bird’s-Eye View Transformation. Sensors 2022, 22, 773. [Google Scholar] [CrossRef] [PubMed]
  9. Cannici, M.; Ciccone, M.; Romanoni, A.; Matteucci, M. Attention mechanisms for object recognition with event-based cameras. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1127–1136. [Google Scholar]
  10. Barranco, F.; Teo, C.L.; Fermuller, C.; Aloimonos, Y. Contour detection and characterization for asynchronous event sensors. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 486–494. [Google Scholar]
  11. Barranco, F.; Fermuller, C.; Ros, E. Real-time clustering and multi-target tracking using event-based sensors. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 5764–5769. [Google Scholar]
  12. Kim, H.; Leutenegger, S.; Davison, A.J. Real-time 3D reconstruction and 6-DoF tracking with an event camera. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 349–364. [Google Scholar]
  13. Rebecq, H.; Ranftl, R.; Koltun, V.; Scaramuzza, D. Events-to-video: Bringing modern computer vision to event cameras. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3857–3866. [Google Scholar]
  14. Scheerlinck, C.; Barnes, N.; Mahony, R. Continuous-time intensity estimation using event cameras. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; pp. 308–324. [Google Scholar]
  15. Wang, L.; Ho, Y.-S.; Yoon, K.-J. Event-based high dynamic range image and very high frame rate video generation using conditional generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10081–10090. [Google Scholar]
  16. Choi, J.; Yoon, K.-J. Learning to super resolve intensity images from events. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2020; pp. 2768–2776. [Google Scholar]
  17. Han, J.; Yang, Y.; Zhou, C.; Xu, C.; Shi, B. Evintsr-net: Event guided multiple latent frames reconstruction and super-resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 4882–4891. [Google Scholar]
  18. Wang, L.; Kim, T.-K.; Yoon, K.-J. Eventsr: From asynchronous events to image reconstruction, restoration, and super-resolution via end-to-end adversarial learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2020; pp. 8315–8325. [Google Scholar]
  19. Gehrig, D.; Loquercio, A.; Derpanis, K.G.; Scaramuzza, D. End-to-end learning of representations for asynchronous event-based data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 15–20 June 2019; pp. 5633–5643. [Google Scholar]
  20. Kim, H.; Handa, A.; Benosman, R.; Ieng, S.-H.; Davison, A.J.J.J.S.S.C. Simultaneous mosaicing and tracking with an event camera. Br. Mach. Vis. Conf. 2008, 43, 566–576. [Google Scholar] [CrossRef]
  21. Cook, M.; Gugelmann, L.; Jug, F.; Krautz, C.; Steger, A. Interacting maps for fast visual interpretation. In Proceedings of the The 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 770–776. [Google Scholar]
  22. Bardow, P.; Davison, A.J.; Leutenegger, S. Simultaneous optical flow and intensity estimation from an event camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 884–892. [Google Scholar]
  23. Munda, G.; Reinbacher, C.; Pock, T. Real-time intensity-image reconstruction for event cameras using manifold regularisation. Int. J. Comput. Vis. 2018, 126, 1381–1393. [Google Scholar] [CrossRef]
  24. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  25. Pan, L.; Scheerlinck, C.; Yu, X.; Hartley, R.; Liu, M.; Dai, Y. Bringing a blurry frame alive at high frame-rate with an event camera. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6820–6829. [Google Scholar]
  26. Pan, L.; Hartley, R.; Scheerlinck, C.; Liu, M.; Yu, X.; Dai, Y. High Frame Rate Video Reconstruction Based on an Event Camera. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2519–2533. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, Z.; Yezzi, A.; Gallego, G. Formulating Event-based Image Reconstruction as a Linear Inverse Problem with Deep Regularization using Optical Flow. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 1, 1–18. [Google Scholar] [CrossRef]
  28. Li, H.; Li, G.; Shi, L. Super-resolution of spatiotemporal event-stream image. Neurocomputing 2019, 335, 206–214. [Google Scholar] [CrossRef]
  29. Duan, P.; Wang, Z.W.; Zhou, X.; Ma, Y.; Shi, B. EventZoom: Learning to denoise and super resolve neuromorphic events. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 12824–12833. [Google Scholar]
  30. Wang, B.; He, J.; Yu, L.; Xia, G.-S.; Yang, W. Event enhanced high-quality image recovery. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 155–171. [Google Scholar]
  31. Wang, Z.W.; Duan, P.; Cossairt, O.; Katsaggelos, A.; Huang, T.; Shi, B. Joint filtering of intensity images and neuromorphic events for high-resolution noise-robust imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2020; pp. 1609–1619. [Google Scholar]
  32. Weng, W.; Zhang, Y.; Xiong, Z. Boosting event stream super-resolution with a recurrent neural network. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part VI. pp. 470–488. [Google Scholar]
  33. Song, C.; Huang, Q.; Bajaj, C. E-cir: Event-enhanced continuous intensity recovery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–22 June 2022; pp. 7803–7812. [Google Scholar]
  34. Gallego, G.; Delbruck, T.; Orchard, G.; Bartolozzi, C.; Taba, B.; Censi, A.; Leutenegger, S.; Davison, A.J.; Conradt, J.; Daniilidis, K.; et al. Event-Based Vision: A Survey. IEEE Trans. Pattern. Anal. Mach. Intell. 2022, 44, 154–180. [Google Scholar] [CrossRef] [PubMed]
  35. Li, Z.; Yang, J.; Liu, Z.; Yang, X.; Jeon, G.; Wu, W. Feedback network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3867–3876. [Google Scholar]
  36. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
  37. Nah, S.; Baik, S.; Hong, S.; Moon, G.; Son, S.; Timofte, R.; Mu Lee, K. Ntire 2019 challenge on video deblurring and super-resolution: Dataset and study. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  38. Rebecq, H.; Gehrig, D.; Scaramuzza, D. ESIM: An open event camera simulator. In Proceedings of the Conference on Robot Learning, Zürich, Switzerland, 29–31 October 2018; pp. 969–982. [Google Scholar]
  39. Diederik, P.K.; Jimmy, B. Adam: A Method for Stochastic Optimization. arXiv 2015, arXiv:1412.6980. [Google Scholar]
  40. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 191–207. [Google Scholar]
  41. Mueggler, E.; Rebecq, H.; Gallego, G.; Delbruck, T.; Scaramuzza, D. The event-camera dataset and simulator: Event-based data for pose estimation, visual odometry, and SLAM. Int. J. Robot. Res. 2017, 36, 142–149. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Three-layer model of a human retina and corresponding event camera pixel circuitry. The first layer is similar to retinal cone cells for photoelectric conversion; the second layer, similar to bipolar cells in the retina, is used to obtain changes in light intensity; the third layer is similar to the ganglion cells of the retina for outputting the light intensity change sign.
Figure 1. Three-layer model of a human retina and corresponding event camera pixel circuitry. The first layer is similar to retinal cone cells for photoelectric conversion; the second layer, similar to bipolar cells in the retina, is used to obtain changes in light intensity; the third layer is similar to the ganglion cells of the retina for outputting the light intensity change sign.
Sensors 23 02155 g001
Figure 2. The process of generating events by the event camera. Each pixel acts as an independent detection unit for luminance changes, and events are generated immediately when the log intensity change at the pixel reaches the specified threshold C t h . Continuous generation of events will form event streams. The event streams contain events of two polarities. When the light intensity changes from strong to weak and reaches the threshold, the camera outputs an OFF event (indicated by the blue arrow); when the light intensity changes from weak to strong and reaches the threshold, the camera outputs an ON event (indicated by the red arrow).
Figure 2. The process of generating events by the event camera. Each pixel acts as an independent detection unit for luminance changes, and events are generated immediately when the log intensity change at the pixel reaches the specified threshold C t h . Continuous generation of events will form event streams. The event streams contain events of two polarities. When the light intensity changes from strong to weak and reaches the threshold, the camera outputs an OFF event (indicated by the blue arrow); when the light intensity changes from weak to strong and reaches the threshold, the camera outputs an ON event (indicated by the red arrow).
Sensors 23 02155 g002
Figure 3. EFSR-Net network structure. The event data is first preprocessed to form a stack, followed by a series of encoding and decoding through the network. The processed event information and APS frame information are used as inputs into the upper and lower coupling sub-networks. Each sub-network consists of a feature extraction block (FEB), a coupled response block (CRB), and a reconstruction block (REB). The final super-resolution image reconstruction is achieved by the mixer (MIX) convolutional network.
Figure 3. EFSR-Net network structure. The event data is first preprocessed to form a stack, followed by a series of encoding and decoding through the network. The processed event information and APS frame information are used as inputs into the upper and lower coupling sub-networks. Each sub-network consists of a feature extraction block (FEB), a coupled response block (CRB), and a reconstruction block (REB). The final super-resolution image reconstruction is achieved by the mixer (MIX) convolutional network.
Sensors 23 02155 g003
Figure 4. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 2× SR on synthetic datasets. The APS frame and event stack are upsampled with bicubic interpolation to the corresponding scale for reference.
Figure 4. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 2× SR on synthetic datasets. The APS frame and event stack are upsampled with bicubic interpolation to the corresponding scale for reference.
Sensors 23 02155 g004
Figure 5. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 4× SR on synthetic datasets. The APS frame and event stack are upsampled with bicubic interpolation to the corresponding scale for reference.
Figure 5. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 4× SR on synthetic datasets. The APS frame and event stack are upsampled with bicubic interpolation to the corresponding scale for reference.
Sensors 23 02155 g005
Figure 6. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 2× SR on real datasets. The APS frame is upsampled with bicubic interpolation to the corresponding scale for reference. The 4× SR results of eSL-Net is downsampled with bicubic interpolation to the corresponding scale for reference.
Figure 6. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 2× SR on real datasets. The APS frame is upsampled with bicubic interpolation to the corresponding scale for reference. The 4× SR results of eSL-Net is downsampled with bicubic interpolation to the corresponding scale for reference.
Sensors 23 02155 g006aSensors 23 02155 g006b
Figure 7. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 4× SR on real datasets. The APS frame and the 2× SR results of E2SRI are upsampled with bicubic interpolation to the corresponding scale for reference.
Figure 7. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 4× SR on real datasets. The APS frame and the 2× SR results of E2SRI are upsampled with bicubic interpolation to the corresponding scale for reference.
Sensors 23 02155 g007
Figure 8. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 2× SR on real datasets. The APS frame is upsampled with bicubic interpolation to the corresponding scale for reference. The 4× SR results of eSL-Net is downsampled with bicubic interpolation to the corresponding scale for reference.
Figure 8. Comparison of the visual quality of our proposed method with other state-of-the-art methods for 2× SR on real datasets. The APS frame is upsampled with bicubic interpolation to the corresponding scale for reference. The 4× SR results of eSL-Net is downsampled with bicubic interpolation to the corresponding scale for reference.
Sensors 23 02155 g008
Figure 9. The effects of different values of N e on event stacks and reconstructed images are qualitatively compared. The APS frame is upsampled with bicubic interpolation to the corresponding scale for reference.
Figure 9. The effects of different values of N e on event stacks and reconstructed images are qualitatively compared. The APS frame is upsampled with bicubic interpolation to the corresponding scale for reference.
Sensors 23 02155 g009
Table 1. Quantitative evaluation of 2× and 4× SR. ↑ means the higher (lower), the better results throughout this paper.
Table 1. Quantitative evaluation of 2× and 4× SR. ↑ means the higher (lower), the better results throughout this paper.
ScaleMethodPSNR ↑SSIM ↑
EV [15] + SISR [40]12.520.466
E2SRI [16]16.410.587
eSL-Net [30]15.760.534
Ours22.020.746
EV [15] + SISR [40]11.930.572
E2SRI [16]--
eSL-Net [30]21.840.683
Ours23.250.714
Table 2. Ablation study of the loss function.
Table 2. Ablation study of the loss function.
LossPSNRSSIM
L121.980.698
L l o s s 22.020.746
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, G.; Feng, Y.; Lv, H.; Zhao, Y.; Liu, H.; Bi, G. Event-Guided Image Super-Resolution Reconstruction. Sensors 2023, 23, 2155. https://doi.org/10.3390/s23042155

AMA Style

Guo G, Feng Y, Lv H, Zhao Y, Liu H, Bi G. Event-Guided Image Super-Resolution Reconstruction. Sensors. 2023; 23(4):2155. https://doi.org/10.3390/s23042155

Chicago/Turabian Style

Guo, Guangsha, Yang Feng, Hengyi Lv, Yuchen Zhao, Hailong Liu, and Guoling Bi. 2023. "Event-Guided Image Super-Resolution Reconstruction" Sensors 23, no. 4: 2155. https://doi.org/10.3390/s23042155

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop