Next Article in Journal
Zero-Trust Model for Smart Manufacturing Industry
Next Article in Special Issue
A Deep Reinforcement Learning Approach for Efficient, Safe and Comfortable Driving
Previous Article in Journal
Analytically Regularized Evaluation of the Coupling of Planar Concentric Conducting Rings
Previous Article in Special Issue
Comprehensive Automated Driving Maneuvers under a Non-Signalized Intersection Adopting Deep Reinforcement Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lightweight Network for Real-Time Rain Streaks and Rain Accumulation Removal from Single Images Captured by AVs

1
School of Engineering, University of Central Lancashire, Preston PR1 2HE, UK
2
Arab Academy for Science, Technology, and Maritime Transport, Alexandria 1029, Egypt
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(1), 219; https://doi.org/10.3390/app13010219
Submission received: 7 December 2022 / Revised: 19 December 2022 / Accepted: 21 December 2022 / Published: 24 December 2022

Abstract

:
In autonomous driving, object detection is considered a base step to many subsequent processes. However, object detection is challenged by loss in visibility caused by rain. Rainfall occurs in two main forms, which are streaks and streaks accumulations. Each degradation type imposes different effect on the captured videos; therefore, they cannot be mitigated in the same way. We propose a lightweight network which mitigates both types of rain degradation in real-time, without negatively affecting the object-detection task. The proposed network consists of two different modules which are used progressively. The first one is a progressive ResNet for rain streaks removal, while the second one is a transmission-guided lightweight network for rain streak accumulation removal. The network has been tested on synthetic and real rainy datasets and has been compared with state-of-the-art (SOTA) networks. Additionally, time performance evaluation has been performed to ensure real-time performance. Finally, the effect of the developed deraining network has been tested on YOLO object-detection network. The proposed network exceeded SOTA by 1.12 dB in PSNR on the average result of multiple synthetic datasets with 2.29× speedup. Finally, it can be observed that the inclusion of different lightweight stages works favorably for real-time applications and could be updated to mitigate different degradation factors such as snow and sun blare.

1. Introduction

Autonomous driving is considered the next-generation transportation breakthrough due to its multiple benefits. According to the World Health Organization (WHO), approximately 1.3 million people die each year as a result of road traffic accidents [1]. Autonomous vehicles (AVs) have the potential to enhance road safety by applying artificial intelligence (AI) techniques in its different tasks [2]. The Society of Automotive Engineers (SAE) ranks autonomous driving from Level 0 (no automation) to Level 5 (full automation under any operational design domains (ODDs), including adverse weather conditions) [3]. Until now, Level 5 AVs have not been introduced to the market; they should be reliable in all challenging conditions in order to gain the public’s confidence that they are worth the investment. Most AV applications are tested in clear weather conditions. However, adverse weather conditions cause sensors (ex: camera, LiDAR, etc.) report inaccurate data and limit their functionality.
Many survey papers have reviewed different algorithms used for object detection in AVs [4,5,6,7,8,9,10,11]. On the other hand, the performance of object-detection algorithms is greatly affected by the presence of challenging weather conditions.
Challenging weather conditions that cause images distortions can be classified into steady and dynamic [12]. Steady weather conditions include fog, mist, and haze. While dynamic weather conditions include rain and snow. Dynamic weather conditions impose greater degradations to the scene’s visibility as the droplets are bigger and of a changing pattern. These conditions reduce scene visibility and contrast, which causes a significant degradation in the process of objects detection performance by AVs.
Rain is the most common dynamic challenging weather condition, especially in the United Kingdom [2]. Even though many researchers have achieved advancements in developing deraining methods, real-time video deraining for AV tasks have been largely understudied.
Rain can be described as the presence of countless drops with different sizes, complex shapes, and varying speeds. Rain causes two different types of visibility degradations. Rain streaks distort the images by generating specular highlights and distorting background scene elements. On the other hand, rain streaks may accumulate generating a hazy effect, similar to fog, hence, reducing the visibility of the scene [13]. Characteristics of rain streaks and rain streaks accumulation are different; therefore, distortions they impose on images are different. Eventually, researchers address them as two separate problems.
In this paper we aim to present a lightweight multi-stage network for single image deraining which mitigates the main two types of rain degradations. Our approach utilizes different progressively stages to reliably restore images affected by different rain degradations while maintaining low inference time. Additionally, extensive experiments are conducted to validate the effectiveness of our network on both synthetic and real rainy datasets. Moreover, despite the existing research towards the development of deraining frameworks, the effect of deraining on object detection was not sufficiently studied especially in the context of autonomous driving; therefore, we also apply object detection for comprehensively evaluating the deraining performance.
The paper is organized as follows. Section 2 we review related deraining work. In Section 3, we introduce and discuss the proposed network. In Section 4, we evaluate the proposed network over different synthetic and real rainy datasets, also we perform sped comparisons between the proposed network and other SOTA networks, experimental results are quantitively and qualitatively presented. In Section 5, we summarized our findings. Moreover, we tabulated used abbreviations in Abbreviations section.

2. Related Work

Deraining problem can be addressed as either a video-based problem or a single image-based one. Video based methods may make use of temporal information present in subsequent frames. On the hand, single image-based deraining is a more challenging task as it lacks the temporal redundancy knowledge and less information is available [14,15]. In order to handle this problem, more research attention has focused on developing different algorithm designs. Moreover, single image-based approaches can be divided into three methods [16]:
  • Filter-based [17,18,19,20]: It is based on identifying different rain properties and designing appropriate filters to achieve rain-free images;
  • Prior-based [21,22,23,24,25,26]: It uses prior knowledge of rainy and clear images. Prior-based methods includes, but not limited to, morphological component analysis (MCA) [27], histogram of oriented gradients (HOGs) [28], structural similarities, and sparse representation models [26];
  • Deep-learning-based models [29,30].
Several approaches have been proposed to achieve rain detection and removal [31,32,33,34,35,36,37]. They have been summarized in Table 1.
However, most of them did not consider AVs system requirements, and suffered from the below listed deficiencies:
  • High computation time [33,35,38];
  • Only addressed the problem of rain detection;
  • Assumes static scenarios [39]; ref. [40]’s approach does not work with highly dynamic raindrops. The approach of [41] fails to separate dynamic textures and moving objects;
  • Not applicable for real-time scenarios:
    • The approach of [31] requires the involvement of a human operator to process the captured frames, also it is bound by certain scenes (lane and building scenes);
    • The approach of [32,33] exploits temporal information during the recovery process;
    • The approach of [34] makes use of previously cleaned frames;
    • The approach of [35] has high computational time;
  • Limited to certain degradation factors: [36,37,38,42].

2.1. Single Image Deraining

Most of the state-of-the-art algorithms are based on convolutional neural networks (CNNs). Different modules can be used in deraining networks, for example:
  • Residual blocks are used in [43] and in [44,45,46,47];
  • Dilated convolution is used in [48,49];
  • Dense blocks are used in [50];
  • Recurrent layers are used in [51,52].
Some approaches used lightweight networks in either a cascaded or Laplacian pyramid manner aiming to enhance the computational efficiency [53,54]; however, this degraded the deraining performance. On the other hand, deep networks are complicated and makes analyzing single modules harder.

2.2. State-of-the-Art Deraining Networks

2.2.1. PRENET

In 2019, Ren et al. [14] provided a simple baseline for deraining networks. They proposed a novel network which repeatedly unfolds a shallow residual network (ResNet). By using a progressive ResNet, they took advantage of recursive computations. Moreover, they introduced recursive layers which exploits dependencies of deep features across stages. Due to its simplicity, this model is suitable as a baseline for future deraining research.

2.2.2. MSPFN

In 2020, Jiang et al. [44] addressed single image rain streak removal. They explored the muti-scale collaborative representation for rain streaks from the perspective of input image scales and hierarchical deep features in a unified framework. Recurrent calculations capture global textures of similar rain streaks at different locations.

2.2.3. MPRNET

In 2021, Zamir et al. [55] developed a multi-stage network, which progressively learns restorations for degraded images; therefore, the whole process is broken down into smaller more manageable sub-processes. In addition to sequentially exchanging information from early stages to late stages; lateral connections between feature processing blocks are also deployed to maintain all information. It works on different image restoration tasks such as denoising, deblurring, and deraining. At earlier stages, encoder–decoders are used to extract multi-scale contextualized features. Later stages operate at the original image resolution to generate spatially accurate outputs. Moreover, they used a supervised attention module (SAM) between every two stages which refines features before passing them to another stage.

2.2.4. HINET

In 2021, HINet was designed by Chen et al. [56] as a simple and powerful multi-stage network. It consists of two subnetworks. Each subnetwork is a U-Net [57]. They introduced a novel block called the half instance normalization block (HIN). The developed network surpasses SOTA on multiple image restoration tasks, such as denoising, deblurring, and deraining.
SOTA deraining networks succeeded in achieving deraining of single images; however, their performance was not evaluated on the performance of object-detection process. Some of them can process different image restoration tasks, and not limited to deraining. Moreover, the running time was not the main concern for the reviewed SOTA networks.
To summarize, many deraining networks and approaches have been developed. However, reviewed approaches suffered from different limitations, such as slow running time, limited to a single rain degradation, or the requirement of a human operator. These limitations disable their employment on AVs. Therefore, the main objective of this proposed research is to develop and evaluate a lightweight deraining network that can mitigate different rain degradations in real-time and enhancing the following object-detection process.

3. Materials and Methods

In Figure 1, we elaborate the proposed deraining framework which involves the following main stages: (i) rain streak removal and (ii) rain accumulation removal. Separate training for both stages has been made as the degradation of rain accumulation is multiplicative and its removal will negatively affect the rain streak removal stage causing larger errors. On the other hand, both stages have been jointly implemented in the testing phase.
When rain streaks accumulate, they generate a veiling effect, similar to the degradation fog imposes on image (see Equation (1)) [58]. Therefore, the rain accumulation removal process is similar to defogging. The first step is performing rain streak removal, followed by rain accumulation removal, and finally reapplying rain streaks removal.
I ( x ) = J ( x )   α ( x ) + A ( 1 α ( x ) ) , J ( x ) = ( B + t r = 1 s s ˜ t r °   R ) ,
where I ( x ) is the observed hazy image at pixel location x ; J is scene radiance; α is the transmission map; A is the global atmospheric light; B is the background layer; s ˜ t r represents a layer of rain streaks having the same direction; t r is the index of rain-streak layers; s is the maximum number of rain-streak layers, and R is a region-dependent variable that indicates the locations of individually visible rain streaks; finally, ° is an element-wise multiplication operation.
Removal of rain accumulation may lead to strengthening the appearance of existing rain streaks. Therefore, in the proposed organization, another rain streak removal stage was re-applied after rain accumulation removal, as the rain accumulation removal stage will cause the rain streaks that may be missed in the first stage of rain streaks removal to be more apparent. The proposed deraining framework is shown in Figure 1.

3.1. Evaluation Metrics

3.1.1. Evaluation of Detection

The detection methods generate bounding boxes to identify the detected region. There are four classification categories, true positive ( T P ), true negative (TN), false positive ( F P ), false negative ( F N ). Precision can be also called positive predictive values (PPVs), as it represents the ratio of positive results that are T P (Equation (2)). Moreover, recall corresponds to the ratio of positive results that are correctly predicted to be positive (Equation (3))
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N ,
Moreover, intersection over union (IoU) is an evaluation metric which is used to compare the predicted bounding box (detection output area) with the ground truth annotated area. It is defined in Equation (4):
I o U = A g t   A P A g t   A P ,
where A g t is the area of the ground-truth of experts annotation and A p is the predicted bounding box from the detetcion method.

3.1.2. Evaluation of Deraining

In order to compare deraining performance, two main assessment indices are used based on reference rain-free images: peak-signal-to-noise ratio and structural similarity index (SSIM) [59]. PSNR is the ratio between the maximum power of a signal and the power of the noise distorting this signal. In order to calculate the PSNR of a test image g and a reference image f . Both images are of size M × N, the PSNR between f and g is defined by Equation (5).
P S N R ( f , g ) = 10   log 10 ( 255 2 M S E ( f , g ) ) ,
where
M S E ( f , g ) = 1 M N i = 1 M   j = 1 N ( f i j g i j ) 2 ,
As the mean square error (MSE) approaches zero, the PSNR value approaches infinity; therefore, high PSNR value corresponds to a better image quality, on the other hand, a small PSNR value implies big numerical differences between images.
Another well-known evaluation metric that is used to measure the similarity between two images is the SSIM. It was developed by Wang et al. [60], It is correlated with the quality perception of the human visual system (HVS). SSIM models image distortion in terms of three factors: loss of correlation, luminance distortion, and contrast distortion. It is defined in Equation (7)
S S I M ( f , g ) = l ( f , g ) c ( f , g ) s ( f , g ) ,  
where   l ( f , g ) corresponds to the luminance comparison, c ( f , g ) corresponds to the contrast comparison, and finally s ( f , g ) corresponds to the structure comparison.
However, in order to evaluate the performance of the deraining process on real rainy images, where their corresponding clear images are not available, we used two other different evaluation metrics: the naturalness image quality evaluator (NIQE) [61] and spatial spectral entropy-based quality (SSEQ) [62].

3.2. Architecture

3.2.1. Rain Streaks Removal

The rain streak removal step is based on residual networks (ResNets) [43]. In order to maintain real-time performance and not adding to the computational complexity of the object-detection framework The base architecture of progressive residual networks has been adopted. So, instead of utilizing deeper and complex networks, rain streak removal is performed in multiple stages, such that ResNet is used at each stage. Additionally, in order to avoid the increase in the number of network parameters and the susceptibility of overfitting, recursive computations have been applied by sharing same network parameters with multiple stages.
A basic ResNet has three parts: (i) a convolutional layer f i n , (ii) multiple ResBlocks f r e s to extract deep representations, and (iii) a convolutional layer f o u t , as shown in Figure 2. Inference of the PRN at stage t is shown in Equation (2).
x t 0.5 = f i n ( x t 1 , y ) , x t = f o u t ( f r e s ( x t 0.5 ) ) ,
Network parameters are reused across different stages, as f i n , f r e s , and f o u t are stage invariant. As shown in Equation (2), f i n takes the concatenation of the current estimation x t 1 and rainy image y as input, the inclusion of y further improves the deraining performance. The progressive network architecture is shown in Figure 3:
  • Filter sizes 3 × 3 , padding 1 × 1 ;
  • f i n is a convolutional layer with ReLU non-linearity [63], due to the concatenation of 3 channels from y and another 3 channels from x t 1 , the convolution of f i n   has 6 input channels and 32 output channels;
  • f r e s includes 5 ResBlocks;
  • f o u t   is a single-layer convolution, it takes the output of f r e s with 32 channels as input and outputs a 3-channel RGB image.

3.2.2. Rain Accumulation Removal

Rain streaks and rain accumulation impose different visual distortion effects on captured images. Distant rain streaks accumulate and impose a distortion similar to fog, which also causes visibility degradation and leads to further hiding of image and present objects’ features, thus should also be removed. Since its distortion effect caused by rain accumulation is similar to fog (Equation (1)), then removal of rain accumulation may follow same algorithm (e.g., [13,58,64]).
The rain accumulation removal network is based on a transmission-guided lightweight neural network (TGL-Net) [65]. In this network, instead of producing simulated transmission maps from both depth data and clear images, guided transmission maps are automatically computed using a filter-refined dark-channel-prior (F-DCP) method which facilitates network training on both synthetic and natural images.
The architecture is based on the very deep residual encoder–decoder network (RED-Net) [66] which owns a symmetric architecture and employs skip connections; however, TGL-Net has fewer layers and parameters and is very lightweight. TGL-Net is composed of three stages: (i) downsampling, (ii) encoder–decoder, and (iii) upsampling, as shown in Figure 4 [65].
The downsampling phase includes a convolutional layer ( 3 × 3 ) then a max pooling layer. The convolutional layer extracts images features, and then the max-pooling layer reduces the total amount by 25 times by using a stride of 5. This leads to improving the computational efficiency of the network.
In the encoder–decoder phase, three convolutional and three deconvolutional layers are connected through a condensed encoder–decoder connection. This stage performs feature extraction and transmission estimation. The convolutional layers extract further image features and also clear noise. On the other hand, the deconvolutional layers recover the details of transmission maps. In the encoder, the sizes of the convolutional kernel are 3 × 3 , 5 × 5 , and 5 × 5 sequentially, which are equal to the corresponding deconvolutional kernels in the decoder, but in reverse.
The symmetric structure of the convolutional and deconvolutional links retains the benefits from the residual network characteristics [43], including rapid convergence and high accuracy. Using skip connections that directly propagate signals to subsequent layers reduces the common problem of vanishing gradients which commonly happens in deep neural networks (DNNs) [67]. Moreover, using skip connections helps maintain useful image features and details.
After the encoder–decoder layers, another convolutional layer is used, it combines a three-channel feature map into a single-channel transmission map. Additionally, a non-linear sigmoid activation function has been applied, it reaches a saturation region when the input is either too small or too big (i.e., larger than 5 or smaller than −5), which increases the risk of vanishing gradient when adding more layers. However, the employment of ELU units and skip connections in the encoder–decoder layers prevent the vanishing gradient problem. The non-linear sigmoid function outputs a preliminary estimation of a reduced transmission map which is fifth the size of the input image.
In order to enlarge the transmission map to the same size as the input image, the upsampling stage is applied. The upsampling stage is composed of two steps. First, image expansion is performed using bilinear interpolation to maintain the computational efficiency of the network. Second, a convolutional layer is used to further refine the output of the transmission map.

4. Experimental Results

4.1. Experimental Setting

Both rain streak removal and rain accumulation removal stages were trained separately. However, they were tested progressively. The rain streak removal network (PRN) was trained on 1800 images of the Rain100H dataset, while the rain accumulation removal network was trained on a synthetic set of hazy images generated from the NYU dataset and augmented real hazy data generated from the NTIRE 2018 dataset. Both networks used the ADAM algorithm optimizer for training with an initial rate of 0.001. The full proposed network was tested and evaluated using a computer equipped with NVIDIA GTX 1050Ti GPU.

4.1.1. Number of Progressive Stages in the Rain Streak Removal Stage

The charts in Figure 5 show that the effect of the number of PRN stages ( T ) on the PSNR and SSIM evaluation metrics on the Rain100H dataset. It can be noted that the more stages in the network, the higher the values of both PSNR and SSIM. However, when T > 8, both PSNR and SSIM stop improving. Additionally, Figure 6, lists the average running time when tested on the Rain100H dataset on models with stages T = 2, 3, 4, 5, 6, 7, 8, 9, and 10 on a computer equipped with NVIDIA GTX 1050Ti GPU. Considering the trade-off between efficiency and deraining performance, we set T   to 8 in the following experiments.

4.1.2. Effect of Integrating Different Stages Progressively

Two experiments were made to correctly choose the right organization of the network’s components. The first suggestion was to apply rain accumulation removal as an initial pre-processing step; however, as shown in Figure 7, this has negatively affected the whole deraining process. This is because the rain accumulation removal step alters the quality of rain streaks, it boosts their contrast and appearance, this makes rain streaks look different compared with their original appearance in both real-world and synthetic datasets. Therefore, the second proposed suggestion was to apply rain accumulation removal progressively in combination with rain streak removal, shown in Figure 8. Additionally, the corresponding PSNR and SSIM values for both suggestions are shown in Table 2.

4.2. Experimental Results

4.2.1. Evaluation on Synthetic Datasets

Our proposed network is evaluated on five synthetic datasets, i.e., Test100 [68], Rain100H [48], Rain100L [48], Test2800 [42], and Test1200 [50]. Both PSNR and SSIM values have been compared and tabulated in Table 3. It can be seen that our proposed model achieves remarkable improvement over SOTA deraining algorithms.
Qualitative results on different datasets, shown in Table 3, are shown in Figure 9. Our proposed model exhibits improved restoration performance especially on Test100, Rain100H, and Test1200 datasets, as these datasets include diverse raining conditions including rain streaks and rain streaks accumulation. Moreover, our proposed network showed equally improved performance on Rain100L and Test2800 datasets. Other networks tend to either blur the images by over-smoothing or leave some rain streaks visible in the images.

4.2.2. Evaluation on Real-World Datasets

Additional experiments have been performed on real-world rainy datasets RID and RIS [69] in order to evaluate the generalization capabilities of our proposed model. Moreover, due to the absence of ground truth (clear with no rain) images, no-reference metrics are required to evaluate the performance of deraining (NIQE and SSEQ). Quantitative results are shown in Table 4.

4.2.3. Time Performance Evaluation

Due to the tight real-time constraints of AVs tasks—in our case, object detection—any additional processing should not impose high computational complexity that may affect the real-time performance. Therefore, deraining process should take this into consideration. Moreover, another challenge imposed by the nature of AVs, is that both Avs and the surrounding objects are in a dynamic/moving state. Since computational efficiency is crucial for autonomous driving which requires real-time performance. Table 5 lists the running time of different methods based on a computer equipped with an NVIDIA GTX 1050Ti GPU. Our proposed model achieves competitive performance compared with other models.
Authors of these SOTA deraining algorithms have evaluated the inference time in their research on different GPUs; therefore, we had to re-evaluate their running times on the same image sizes on a unified computer to produce valid comparisons.

4.2.4. Effect on Object-Detection Performance

Deraining of images captured by AVs in real driving scenarios is considered an effective enhancement of images quality which subsequently enhances the performance of object-detection task performed by AVs, which is the main objective of our work. Therefore, we investigated the effect of applying deraining algorithms, including our proposed network, on the accuracy of object detection.
In order to perform this evaluation, we used samples collected by Jian et al. in [44]. They have collected samples from datasets that include driving scenes with diverse conditions (850 samples from COCO and BDD datasets to create new synthetic rain datasets: COCO350, and BDD350), such as diverse rain orientations and magnitudes, in addition to complex scenes (e.g., night scenes). First, we applied our proposed deraining algorithm, as well as other SOTA deraining algorithms in order to restore rain-free images. Afterwards, we apply the object-detection framework—YOLOv3—to assess the detection efficiency on de-rained images. Quantitative results of both the deraining performance as well as the detection efficiency are shown in Table 6. In addition, visual comparisons are shown in Figure 10.
The object-detection accuracy is greatly degraded by the presence of unclear weather conditions, especially rain. Rain drops cover and distort underlying visual features of surrounding objects, which are used by detection methods to classify and localize objects. Figure 10 demonstrates the performance of object detection with and without draining in advance.
Not all deraining algorithms boosts the performance of object detection; however, Hnewa et al. in [70] and Li et al. in [69] have shown that some deraining algorithms [43,51] may degrade the detection performance when compared with directly using the rainy images as input to the corresponding detection frameworks, as they are not optimized towards the end goal of object detection. One factor for this degradation is that the tested deraining algorithms tended to smooth out the images, and hence, distorts important visual features of a scene, and smooths out edges of objects leading to missing detecting them.
Therefore, in our research it is crucial to evaluate the performance of our proposed deraining model by jointly evaluating the object-detection performance on the de-rained output. As shown in Table 6, the detection precision of the produced de-rained images by our proposed model shows noticeable improvement compared with that of original rainy inputs by 68%. Additionally, when compared with other SOTA deraining models, the de-rained images generated by our proposed model shows more dependable contents with clearer features, which efficiently boosts the detection performance.

5. Conclusions

A novel deraining network is presented; it mitigates different rain effects—rain streaks with different directions and intensities and rain accumulations—and targets images captured under rainy driving scenes. The proposed network comprises of three main phases. The first phase performs rain streak removal to remove most visible rain streaks in an image; the second phase is rain accumulation removal which is similar to a defogging operation; lastly, we perform a second rain streak removal operation in order to remove any left rain streaks that could be boosted by the previous rain accumulation removal phase.
The network is based on baseline lightweight network in order to maintain real-time computational performance, as this is a crucial constraint for the performance of deraining networks along with other subsequent operations performed by AVs.
Experiments showed that the proposed network has higher efficiency when compared with SOTA deraining networks, regarding both qualitative and quantitative evaluation metrics, and running time. Moreover, we evaluate the proposed network in addition to the SOTA deraining networks regarding object-detection performance. The presented methodology is considered to be the first to deal with both real-time efficiencies and enhanced performance.
Further research direction includes the investigation of enhancing the model to mitigate more diverse weather conditions such as snow, sun glare, and lightning. Moreover, the proposed network performs removal of both rain streaks and rain streaks accumulation even if only one of these degradations exist. Therefore, another direction to our research is to add initial step of detecting the type of degradation exists in the captured images. Additionally, we plan to test the network on different hardware in order to evaluate the speedup enhancement that could be achieved using more powerful hardware.

Author Contributions

Conceptualization, E.K. and A.O.; data curation, E.K.; formal analysis, E.K.; funding acquisition, A.O. and M.V.; investigation, E.K.; methodology, E.K.; project administration, A.O. and M.V.; resources, A.O.; software, E.K.; supervision, A.O.; validation, E.K., A.O. and M.V.; visualization, A.A.; writing—original draft, E.K.; writing—review and editing, A.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AbbreviationFull Description
AIArtificial Intelligence
AVAutonomous Vehicle
HINetHalf Instance Normalization Network
MPRNetMulti-Stage Progressive Image Restoration Network
MSPFNMulti-Scale Progressive Fusion Network
NIQENatural Image Quality Evaluator
ODDOperational Design Domain
PReNetProgressive Recurrent Residual Network
PSNRPeak Signal to Noise Ratio
ResNetResidual Network
SAESociety of Automotive Engineers
SOTAState of the art
SSEQSpatial-Spectral Entropy-based Quality
SSIMStructural Similarity Index Measure
WHOWorld Health Organization

References

  1. Road Traffic Injuries. Available online: https://www.who.int/news-room/fact-sheets/detail/road-traffic-injuries (accessed on 30 June 2022).
  2. Yoneda, K.; Suganuma, N.; Yanase, R.; Aldibaja, M. Automated driving recognition technologies for adverse weather conditions. IATSS Res. 2019, 43, 253–262. [Google Scholar] [CrossRef]
  3. SAE. International: On-road automated vehicle standards committee and others. In Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles; SAE International: Warrendale, PA, USA, 2018. [Google Scholar]
  4. Zhao, Z.; Zheng, P.; Xu, S.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Khatab, E.; Onsy, A.; Varley, M. Vulnerable objects detection for autonomous driving: A review. Integration 2021, 78, 36–48. [Google Scholar] [CrossRef]
  6. Zhu, H.; Yuen, K.; Mihaylova, L.; Leung, H. Overview of environment perception for intelligent vehicles. IEEE Trans. Intell. Transp. Syst. 2017, 18, 2584–2601. [Google Scholar] [CrossRef] [Green Version]
  7. Van Brummelen, J.; O’Brien, M.; Gruyer, D.; Najjaran, H. Autonomous vehicle perception: The technology of today and tomorrow. Transp. Res. Part C Emerg. Technol. 2018, 89, 384–406. [Google Scholar] [CrossRef]
  8. Mukhtar, A.; Xia, L.; Tang, T.B. Vehicle detection techniques for collision avoidance systems: A review. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2318–2338. [Google Scholar] [CrossRef]
  9. Sharma, K.U.; Thakur, N.V. A review and an approach for object detection in images. Int. J. Comput. Vis. Robot. 2017, 7, 196–237. [Google Scholar] [CrossRef]
  10. Tiwari, M.; Singhai, R. A review of detection and tracking of object from image and video sequences. Int. J. Comput. Intell. Res. 2017, 13, 745–765. [Google Scholar]
  11. Zhiqiang, W.; Jun, L. A review of object detection based on convolutional neural network. In Proceedings of the 2017 36th Chinese Control Conference (CCC), Dalian, China, 26–28 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 11104–11109. [Google Scholar]
  12. Garg, K.; Nayar, S.K. Vision and rain. Int. J. Comput. Vis. 2007, 75, 3–27. [Google Scholar] [CrossRef]
  13. Yang, W.; Tan, R.T.; Feng, J.; Guo, Z.; Yan, S.; Liu, J. Joint rain detection and removal from a single image with contextualized deep networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1377–1393. [Google Scholar] [CrossRef]
  14. Ren, D.; Zuo, W.; Hu, Q.; Zhu, P.; Meng, D. Progressive image deraining networks: A better and simpler baseline. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3937–3946. [Google Scholar]
  15. Wang, T.; Yang, X.; Xu, K.; Chen, S.; Zhang, Q.; Lau, R.W. Spatial Attentive Single-Image Deraining with a High Quality Real Rain Dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12270–12279. [Google Scholar]
  16. Wang, H.; Wu, Y.; Li, M.; Zhao, Q.; Meng, D. A survey on rain removal from video and single image. arXiv 2019, arXiv:1909.08326. [Google Scholar] [CrossRef]
  17. Xu, J.; Zhao, W.; Liu, P.; Tang, X. Removing rain and snow in a single image using guided filter. In Proceedings of the 2012 IEEE International Conference on Computer Science and Automation Engineering (CSAE), Zhangjiajie, China, 25–27 May 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 304–307. [Google Scholar]
  18. Zheng, X.; Liao, Y.; Guo, W.; Fu, X.; Ding, X. Single-image-based rain and snow removal using multi-guided filter. In Proceedings of the International Conference on Neural Information Processing, Lake Tahoe, NV, USA, 5–10 December 2013; Springer: Berlin, Germany, 2013; pp. 258–265. [Google Scholar]
  19. Ding, X.; Chen, L.; Zheng, X.; Huang, Y.; Zeng, D. Single image rain and snow removal via guided L0 smoothing filter. Multimed. Tools Appl. 2016, 75, 2697–2712. [Google Scholar] [CrossRef]
  20. Kim, J.; Lee, C.; Sim, J.; Kim, C. Single-image deraining using an adaptive nonlocal means filter. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 914–917. [Google Scholar]
  21. Fu, Y.; Kang, L.; Lin, C.; Hsu, C. Single-frame-based rain removal via image decomposition. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1453–1456. [Google Scholar]
  22. Chen, D.; Chen, C.; Kang, L. Visual depth guided color image rain streaks removal using sparse coding. IEEE Trans. Circuits Syst. Video Technol. 2014, 24, 1430–1455. [Google Scholar] [CrossRef]
  23. Kang, L.; Lin, C.; Lin, C.; Lin, Y. Self-learning-based rain streak removal for image/video. In Proceedings of the 2012 IEEE International Symposium on Circuits and Systems (ISCAS), Seoul, Republic of Korea, 20–23 May 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1871–1874. [Google Scholar]
  24. Wang, Y.; Liu, S.; Chen, C.; Zeng, B. A hierarchical approach for rain or snow removing in a single color image. IEEE Trans. Image Process. 2017, 26, 3936–3950. [Google Scholar] [CrossRef] [PubMed]
  25. Sun, S.; Fan, S.; Wang, Y.F. Exploiting image structural similarity for single image rain removal. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 4482–4486. [Google Scholar]
  26. Gu, S.; Meng, D.; Zuo, W.; Zhang, L. Joint convolutional analysis and synthesis sparse representation for single image layer separation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1708–1716. [Google Scholar]
  27. Starck, J.; Moudden, Y.; Bobin, J.; Elad, M.; Donoho, D.L. Morphological component analysis, Wavelets XI. In Proceedings of the Optics and Photonics Conference 2005, San Diego, CA, USA, 31 July–4 August 2005; SPIE: Bellingham, WA, USA, 2005; pp. 209–223. [Google Scholar]
  28. Zhao, X.; Liu, P.; Liu, J.; Xianglong, T. The application of histogram on rain detection in video. In Proceedings of the 11th Joint International Conference on Information Sciences, Khulna, Bangladesh, 24–27 December 2008; Atlantis Press: Paris, France, 2008; pp. 382–387. [Google Scholar]
  29. Wang, M.; Mai, J.; Cai, R.; Liang, Y.; Wan, H. Single image deraining using deep convolutional networks. Multimed. Tools Appl. 2018, 77, 25905–25918. [Google Scholar] [CrossRef]
  30. Wang, X.; Li, Z.; Shan, H.; Tian, Z.; Ren, Y.; Zhou, W. Fastderainnet: A deep learning algorithm for single image deraining. IEEE Access. 2020, 8, 127622–127630. [Google Scholar] [CrossRef]
  31. Liao, H.; Wang, D.; Yang, C.; Shine, J. Video-based water drop detection and removal method for a moving vehicle. Inf. Technol. J. 2013, 12, 569–583. [Google Scholar] [CrossRef] [Green Version]
  32. Chen, J.; Chau, L. A rain pixel recovery algorithm for videos with highly dynamic scenes. IEEE Trans. Image Process. 2013, 23, 1097–1104. [Google Scholar] [CrossRef]
  33. Fu, X.; Huang, J.; Ding, X.; Liao, Y.; Paisley, J. Clearing the skies: A deep network architecture for single-image rain removal. IEEE Trans. Image Process. 2017, 26, 2944–2956. [Google Scholar] [CrossRef] [Green Version]
  34. Yang, W.; Liu, J.; Feng, J. Frame-Consistent Recurrent Video Deraining With Dual-Level Flow. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1661–1670. [Google Scholar]
  35. Li, M.; Xie, Q.; Zhao, Q.; Wei, W.; Gu, S.; Tao, J.; Meng, D. Video rain streak removal by multiscale convolutional sparse coding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6644–6653. [Google Scholar]
  36. Yeh, C.; Lin, C.; Muchtar, K.; Liu, P. Rain streak removal based on non-negative matrix factorization. Multimed. Tools Appl. 2018, 77, 20001–20020. [Google Scholar] [CrossRef]
  37. Chen, J.; Tan, C.; Hou, J.; Chau, L.; Li, H. Robust video content alignment and compensation for rain removal in a cnn framework. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6286–6295. [Google Scholar]
  38. Wang, C.; Shen, M.; Yao, C. Rain streak removal by multi-frame-based anisotropic filtering. Multimed. Tools Appl. 2017, 76, 2019–2038. [Google Scholar] [CrossRef]
  39. Krishnan, S.; Venkataraman, D. Restoration of video by removing rain. Int. J. Comput. Sci. Eng. Appl. 2012, 2, 19. [Google Scholar] [CrossRef]
  40. You, S.; Tan, R.T.; Kawakami, R.; Ikeuchi, K. Adherent raindrop detection and removal in video. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1035–1042. [Google Scholar]
  41. Ren, W.; Tian, J.; Han, Z.; Chan, A.; Tang, Y. Video desnowing and deraining based on matrix decomposition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4210–4219. [Google Scholar]
  42. Fu, X.; Huang, J.; Zeng, D.; Huang, Y.; Ding, X.; Paisley, J. Removing rain from single images via a deep detail network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3855–3863. [Google Scholar]
  43. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  44. Jiang, K.; Wang, Z.; Yi, P.; Chen, C.; Huang, B.; Luo, Y.; Ma, J.; Jiang, J. Multi-scale progressive fusion network for single image deraining. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8346–8355. [Google Scholar]
  45. Abdelhamed, A.; Timofte, R.; Brown, M.S. Ntire 2019 challenge on real image denoising: Methods and results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  46. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C. Deep learning on image denoising: An overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef] [PubMed]
  47. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Yang, W.; Tan, R.T.; Feng, J.; Liu, J.; Guo, Z.; Yan, S. Deep joint rain detection and removal from a single image. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1357–1366. [Google Scholar]
  49. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  50. Zhang, H.; Patel, V.M. Density-aware single image de-raining using a multi-stream dense network. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 695–704. [Google Scholar]
  51. Qian, R.; Tan, R.T.; Yang, W.; Su, J.; Liu, J. Attentive generative adversarial network for raindrop removal from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2482–2491. [Google Scholar]
  52. Li, X.; Wu, J.; Lin, Z.; Liu, H.; Zha, H. Recurrent squeeze-and-excitation context aggregation net for single image deraining. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 254–269. [Google Scholar]
  53. Fu, X.; Liang, B.; Huang, Y.; Ding, X.; Paisley, J. Lightweight pyramid networks for image deraining. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 1794–1807. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  54. Fan, Z.; Wu, H.; Fu, X.; Hunag, Y.; Ding, X. Residual-guide feature fusion network for single image deraining. arXiv arXiv:1804.07493, 2018.
  55. Waqas Zamir, S.; Arora, A.; Khan, S.; Hayat, M.; Shahbaz Khan, F.; Yang, M.; Shao, L. Multi-Stage Progressive Image Restoration. arXiv 2021, arXiv:2102.02808. [Google Scholar]
  56. Chen, L.; Lu, X.; Zhang, J.; Chu, X.; Chen, C. HINet: Half instance normalization network for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 182–192. [Google Scholar]
  57. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin, Germany, 2015; pp. 234–241. [Google Scholar]
  58. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An end-to-end system for single image haze removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [Green Version]
  59. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Washington, DC, USA, 23–26 August 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 2366–2369. [Google Scholar]
  60. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  61. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal. Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  62. Liu, L.; Liu, B.; Huang, H.; Bovik, A.C. No-reference image quality assessment based on spatial and spectral entropies. Signal. Process. Image Commun. 2014, 29, 856–863. [Google Scholar] [CrossRef]
  63. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010. [Google Scholar]
  64. Narasimhan, S.G.; Nayar, S.K. Vision and the atmosphere. Int. J. Comput. Vis. 2002, 48, 233–254. [Google Scholar] [CrossRef]
  65. Li, Z.; Zhang, J.; Zhong, R.; Bhanu, B.; Chen, Y.; Zhang, Q.; Tang, H. Lightweight and Efficient Image Dehazing Network Guided by Transmission Estimation from Real-World Hazy Scenes. Sensors 2021, 21, 960. [Google Scholar] [CrossRef]
  66. Mao, X.; Shen, C.; Yang, Y. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2016; Volume 29. [Google Scholar]
  67. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the 13th International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  68. Zhang, H.; Sindagi, V.; Patel, V.M. Image de-raining using a conditional generative adversarial network. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 3943–3956. [Google Scholar] [CrossRef]
  69. Li, S.; Araujo, I.B.; Ren, W.; Wang, Z.; Tokuda, E.K.; Junior, R.H.; Cesar-Junior, R.; Zhang, J.; Guo, X.; Cao, X. Single image deraining: A comprehensive benchmark analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3838–3847. [Google Scholar]
  70. Hnewa, M.; Radha, H. Object detection under rainy conditions for autonomous vehicles. arXiv 2020, arXiv:2006.16471. [Google Scholar]
Figure 1. Proposed deraining framework.
Figure 1. Proposed deraining framework.
Applsci 13 00219 g001
Figure 2. A basic ResNet.
Figure 2. A basic ResNet.
Applsci 13 00219 g002
Figure 3. Progressive network architecture.
Figure 3. Progressive network architecture.
Applsci 13 00219 g003
Figure 4. TGL-Net’s Symmetric architecture, comprised of three stages: downsampling, encoder–decoder, and upsampling. Dotted lines are skip connections [65].
Figure 4. TGL-Net’s Symmetric architecture, comprised of three stages: downsampling, encoder–decoder, and upsampling. Dotted lines are skip connections [65].
Applsci 13 00219 g004
Figure 5. Effect of increasing number of progressive stages in rain streak removal on PSNR and SSIM evaluation metrics.
Figure 5. Effect of increasing number of progressive stages in rain streak removal on PSNR and SSIM evaluation metrics.
Applsci 13 00219 g005
Figure 6. Effect of increasing number of progressive stages in rain streak removal network on running time (in seconds).
Figure 6. Effect of increasing number of progressive stages in rain streak removal network on running time (in seconds).
Applsci 13 00219 g006
Figure 7. The effect of applying rain accumulation removal as a pre-processing step, and then rain streak removal.
Figure 7. The effect of applying rain accumulation removal as a pre-processing step, and then rain streak removal.
Applsci 13 00219 g007
Figure 8. The effect of performing rain streak removal and rain accumulation removal progressively.
Figure 8. The effect of performing rain streak removal and rain accumulation removal progressively.
Applsci 13 00219 g008
Figure 9. Qualitative results on synthesized datasets.
Figure 9. Qualitative results on synthesized datasets.
Applsci 13 00219 g009
Figure 10. Examples of joint deraining and object detection. The first row denotes samples of the deraining results on COCO350 dataset; the second row denotes samples of the deraining results on BDD350 dataset. The third and fourth rows shows the results of object detection performed on the first and second rows sequentially. YOLOv3 has been used for object detection.
Figure 10. Examples of joint deraining and object detection. The first row denotes samples of the deraining results on COCO350 dataset; the second row denotes samples of the deraining results on BDD350 dataset. The third and fourth rows shows the results of object detection performed on the first and second rows sequentially. YOLOv3 has been used for object detection.
Applsci 13 00219 g010
Table 1. Rain detection and removal approaches.
Table 1. Rain detection and removal approaches.
ApproachImage/
Video
Addressed DegradationMethodLimitation
Liao et al. [31]VideoRain dropsThey used Hough transformation to detect raindrops in the context of lane scenes, and Sobel filter to detect raindrops in the context of a building sceneIt requires the interaction of a human operator to process images
Chen et al. [32]VideoRain streaksThey exploited both spatial and temporal information. For dynamic scenes, they applied motion segmentation. Rain removal filters are applied after photometric and chromatic constraints are used for rain detectionDepends on temporal information
Fu et al. [33]ImageRain streaksDe-rainNet: They used image processing domain knowledge to improve deraining with a modestly sized CNNLow speed
Yang et al. [34]VideoRain streaks and accumulationA rain-free frame is estimated from a single rain frame and then taken as a guidance along with previously recovered clean frames Makes use of previously cleared frames to obtain a more accurate clean one
Li et al. [35]Video Rain streaksApplied a multiscale convolutional sparse coding for rain removalLow Speed
Yeh et al. [36]Image Rain streaksFrequency division, non-negative matrix factorizationLimited to rain streaks
Chen et al. [37]VideoRain streaksThey applied super pixel segmentation to perform scene decomposition into depth consistent unitsLimited to rain streaks.
Table 2. Effect of applying rain accumulation removal as pre-processing vs. applying it progressively.
Table 2. Effect of applying rain accumulation removal as pre-processing vs. applying it progressively.
ApproachPSNRSSIM
Rain accumulation removal as an initial stage15.6830.683
Rain accumulation removal performed progressively with rain streak removal33.010.901
Table 3. Average PSNR and SSIM comparison on the synthetic datasets Test100, Rain100H, Rain100L, Test2800, and Test1200.
Table 3. Average PSNR and SSIM comparison on the synthetic datasets Test100, Rain100H, Rain100L, Test2800, and Test1200.
MethodTest100 [68]Rain100H [48]Rain100L [48]Test2800 [42]Test1200 [50]Average
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
HINet [56]30.290.90630.650.89437.280.97033.910.94133.050.91933.0360.926
MPRNet [55]30.270.89730.410.89036.400.96533.640.93832.910.91632.7260.921
MSPFN [44]27.050.87628.660.86032.400.93332.820.93032.390.91630.7260.903
PreNet [14]24.810.85126.770.85832.440.95031.750.91631.360.91129.4260.897
PRN [14]23.510.76128.070.88436.980.97723.790.81219.730.70926.4170.829
Proposed32.430.93033.010.90138.210.96933.890.93234.230.95334.1540.937
Table 4. Comparison results of average NIQE/SSEQ on real-world dataset (RID and RIS). The smaller scores indicate better perceptual quality.
Table 4. Comparison results of average NIQE/SSEQ on real-world dataset (RID and RIS). The smaller scores indicate better perceptual quality.
DatasetEvaluation MetricsHINetMPRNetMSPFNPreNetProposed
RIDNIQE ↓4.9854.8566.5187.0074.562
SSEQ ↓81.16250.64840.4743.0448.213
RISNIQE ↓6.8876.0456.1356.7226.148
SSEQ ↓52.98351.68943.4748.2252.681
Table 5. Comparison of running time (sec) on NVIDIA GTX 1050Ti GPU.
Table 5. Comparison of running time (sec) on NVIDIA GTX 1050Ti GPU.
Image SizeHINetMPRNetMSPFNPreNetProposed
500 × 5000.5740.7920.4150.4640.251
1024 × 10242.0652.7821.5381.5780.843
Table 6. Comparison results of joint image deraining and object detection on samples of COCO and BDD datasets.
Table 6. Comparison results of joint image deraining and object detection on samples of COCO and BDD datasets.
MethodsRainy InputHINetMPRNetMSPFNPreNetProposed
Deraining; Dataset COCO350/BDD350
PSNR14.79/14.1318.068/15.6118.01/16.6518.23/17.8517.53/16.918.879/17.98
SSIM0.648/0.4700.786/0.5670.786/0.7790.782/0.7610.765/0.6520.801/0.823
Object Detection; Algorithm: YOLOv3; Dataset: COCO350/BDD350
Precision (%)23.03/36.8632.87/41.5231.45/40.0532.56/41.0431.31/38.6633.54/41.97
Recall (%)29.6/42.839.54/50.2138.21/49.3239.31/50.4037.92/48.5940.12/50.67
IoU (%)55.5/59.8560.57/62.5362.54/62.2162.54/62.2160.75/61.0862.34/61.23
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khatab, E.; Onsy, A.; Varley, M.; Abouelfarag, A. A Lightweight Network for Real-Time Rain Streaks and Rain Accumulation Removal from Single Images Captured by AVs. Appl. Sci. 2023, 13, 219. https://doi.org/10.3390/app13010219

AMA Style

Khatab E, Onsy A, Varley M, Abouelfarag A. A Lightweight Network for Real-Time Rain Streaks and Rain Accumulation Removal from Single Images Captured by AVs. Applied Sciences. 2023; 13(1):219. https://doi.org/10.3390/app13010219

Chicago/Turabian Style

Khatab, Esraa, Ahmed Onsy, Martin Varley, and Ahmed Abouelfarag. 2023. "A Lightweight Network for Real-Time Rain Streaks and Rain Accumulation Removal from Single Images Captured by AVs" Applied Sciences 13, no. 1: 219. https://doi.org/10.3390/app13010219

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop