Next Article in Journal
An On-Site InSAR Terrain Imaging Method with Unmanned Aerial Vehicles
Previous Article in Journal
Impact of Reducing Statistically Small Population Sampling on Threshold Detection in FBG Optical Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

S-LIGHT: Synthetic Dataset for the Separation of Diffuse and Specular Reflection Images

Advanced Visual Intelligence Lab (AVILAB), Yeungnam University, Gyeongsan-si 38541, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(7), 2286; https://doi.org/10.3390/s24072286
Submission received: 10 February 2024 / Revised: 19 March 2024 / Accepted: 1 April 2024 / Published: 3 April 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Several studies in computer vision have examined specular removal, which is crucial for object detection and recognition. This research has traditionally been divided into two tasks: specular highlight removal, which focuses on removing specular highlights on object surfaces, and reflection removal, which deals with specular reflections occurring on glass surfaces. In reality, however, both types of specular effects often coexist, making it a fundamental challenge that has not been adequately addressed. Recognizing the necessity of integrating specular components handled in both tasks, we constructed a specular-light (S-Light) DB for training single-image-based deep learning models. Moreover, considering the absence of benchmark datasets for quantitative evaluation, the multi-scale normalized cross correlation (MS-NCC) metric, which considers the correlation between specular and diffuse components, was introduced to assess the learning outcomes.

1. Introduction

Research in computer vision has utilized various types of image data, such as electric optical (EO), infrared (IR), and light detection and ranging (Lidar). Among these, EO images, which are within the visible spectrum perceivable by humans (380 nm to 780 nm), have been studied extensively. EO images have been used for research object detection, classification, image registration, 3D transformation, and other areas, due to their unique characteristics and patterns. The advent of deep learning has prompted significant growth in EO image-related research. On the other hand, deep learning models based solely on single EO images are not robust enough to perform well in various environments. Recent deep-learning research has addressed this issue by developing more robust models that combine sensors from EO and other modalities [1,2].
One of the causes of the drawbacks associated with EO images is the presence of specular reflection components. The term “specular reflection component” refers to the portion of light energy that impinges on the surface of an object and is reflected without change, which can degrade the performance of object classification, detection, and segmentation in computer vision. Efforts to address and eliminate these reflection components have been ongoing in computer vision research since its inception, and research to solve this problem continues to the present day. Specular reflection components vary in their characteristics depending on the material, which has led to research in two main categories: reflection removal and specular highlight removal.
Specular highlight removal: The problem in specular highlight removal can be described using Equation (1). Figure 1 provides an example image representing the diffuse and specular components occurring on the surface of an object. In Equation (1), the diffuse component is denoted as I d , the specular component as I s , and the combined intensity as I.
I = I d + I s
Here, the diffuse component refers to the light rays that penetrate the surface of an object, undergo multiple reflections and refractions, and are then re-emitted. On the other hand, specular refers to the portion of incoming light rays that are reflected directly, and its intensity varies depending on the roughness of the surface of an object.
Reflection removal: Unlike the problem tackled by specular highlight removal, the primary goal of the reflection removal task is to eliminate reflections occurring in situations with a medium, such as glass, between the object and the viewer and to generate an image of the object that has passed through the glass. In Figure 2, transmission and reflection represent I t and I r in Equation (2).
I = I t + I r
Each component undergoes transformations depending on the properties of the glass. The variable I t is in a state where it has experienced pixel shifts and color alterations due to the refraction and absorption effects of the glass, whereas I r is affected by reflection and refraction effects in Equation (2). Taking these physical characteristics into consideration, most techniques model I t and I r in Equation (2) as follows.
I t = α I T
I r = β ( I R k )
In Equation (3), I T represents the image before passing through the glass, α denotes the weighting coefficient, and in Equation (4), β also represents the weighting coefficient. The image I R before being reflected by the glass undergoes convolution with the potential degradation kernel k.
The reflection removal task focuses on removing images reflected on glass surfaces to produce transmission images. In contrast, specular highlight removal aims to eliminate the specular component images on inhomogeneous object surfaces to generate the remaining image.
Research has been conducted using single and multiple images to address these two categories. In the early stages of research, the problem of removing specular reflections from a single image was considered an ill-posed problem, leading to studies that relied heavily on handcrafted priors. Using a single image allows easy data collection and high computational efficiency, but it is not robust under various environmental conditions. Researchers have also explored the use of multiple images to overcome the performance limitations of single images. These approaches include methods [3,4,5,6] based on multi-view images that leverage the viewpoint-dependent nature of the specular component. The approaches also include methods [7,8,9,10,11] based on multiple polarized filter images that exploit polarization when light reflects from object surfaces and some methods [12,13] based on flash and non-flash image pairs. Although research based on multiple images offered better performance and robustness than single-image techniques, there are practical limitations because of the challenges of acquiring multiple images in real-world settings and the high computational demands.
Recent advances in deep learning techniques have led to a significant breakthrough in removing specular highlights and reflection components from single images. Despite these advances, research on these two tasks is still being conducted independently, and they are not robust against each other’s tasks, which limits their practicality. Figure 3 displays images containing specular reflection components frequently encountered in everyday life.
In the real world, specular highlight and reflection do not occur selectively. Hence, we advocate for the need for research into unified specular reflection removal that tackles both tasks. However, there are major challenges in driving unified reflection removal research, and to the best of our knowledge, there is no relevant research to date. Acquiring diffuse and specular reflection images in the uncontrollable conditions of the real world is nearly impossible. This issue not only prevents the use of real images as a training dataset but also precludes the possibility of conducting quantitative evaluations.
To alleviate this issue, this paper proposes a training dataset for single image-based deep learning models that considers recent developments in deep learning techniques and practical considerations. The paper also discusses which category of deep learning models is suitable for training and demonstrates their effectiveness in real-world images. The contributions are as follows:
  • First, a synthetic dataset was constructed for training a single-image-based deep learning model that can consider both the specular components reflected on the surface and those reflected on the glass, a consideration not present in traditional specular highlight and reflection removal.
  • Second, this paper proposes a performance metric that considers the relationship between the diffuse component and each specular component to measure the separation level quantitatively. Additionally, it offers the advantage of being able to visualize the results.
This paper is structured as follows. Section 2 covers non-learning and deep-learning methods for specular highlight removal and reflection removal tasks and the datasets used for training deep-learning models. Section 3 addresses the dataset that considers integrated specular components and the performance metric that takes the correlation between the diffuse component and the integrated specular component into account. Section 4 presents experimental comparisons of the proposed dataset and existing datasets regarding the learning performance and generalization, along with the results of applying the methods to real-world images. Finally, Section 5 reports the conclusion and outlines future research directions.
The energy of wavelengths in the visible spectrum undergoes numerous interactions before reaching the camera and the human eye. Modeling these phenomena remains a critical research topic in computer graphics and computer vision. Section 2 overviews how previous studies tackled these two categories. The section is divided into two parts, with the first and second parts discussing recent trends and the limitations of existing methods, respectively.

2. Related Work

This section comprises three subsections, each covering the approach taken for the respective tasks. First, Section 2.1 describes non-deep learning (Non-DL) techniques; second, Section 2.2 discusses deep learning (DL) techniques; and finally, in order to prepare a dataset for the deep learning models, Section 2.3 details the methods for obtaining training data.

2.1. Non-Deep Learning (Non-DL)

Specular Highlight Removal: Removing specular highlights has been a subject of research for a long time. Shafer [14] proposed a dichromatic reflection model to examine the components present on the surface of an object. This model was formulated assuming the object was opaque and inhomogeneous. The dichromatic reflection model has been widely cited and has significantly influenced the field of computer graphics and subsequent research in specular highlight removal. Klinker et al. [15] extended the concepts of the research [14] by considering that the distribution of pixels on the plane follows a t-shaped color distribution. They proposed an algorithm based on this concept. Bajcsy et al. [16] suggested a technique that performs this task in a color domain consisting of lightness, saturation, and hue. On the other hand, these techniques require color segmentation techniques, which can be a significant drawback in complex textures.
Tan and Ikeuchi [17] aimed to overcome these issues by proposing an effective method that can remove specular highlights without the need for explicit color segmentation. They introduced the specular-to-diffuse mechanism and the concept of a specular-free image containing only the diffuse reflection component. Based on the observation that the specular-free image has the same geometric distribution as the original intensity image, they devised a method to estimate the diffuse reflection through the logarithmic differentiation of the original and specular-free images. This approach significantly influenced further research [18,19,20,21].
Subsequent research focused heavily on optimization techniques. Kim et al. [22] proposed a maximum a posteriori (MAP) optimization technique based on the observation that the dark channel provides pseudo-specular reflection results in general natural images. Fu et al. [23] used k-means clustering and various priors in an optimization technique, and Akashi and Okatani [24] proposed an optimization technique based on sparse non-negative Matrix factorization (NMF).
Reflection Removal: The work of Levin et al. [25] is the first proposed method for separating mixed images with reflections using a single image. It introduced an optimization algorithm that minimizes the total amount of edges and corners in the separated images through a prior that they should be minimal. On the other hand, this algorithm was not robust in handling complex pattern images. Removing reflections using a single image is highly ill-posed.
Various studies have been conducted based on different priors and assumptions to make it more tractable. Levin and Weiss [26] proposed a method that utilizes gradient sparsity priors after users specify a few labels on the image. Li and Brown [27] used the fact that the camera focus is adjusted for a transmission image, proposing an optimization technique that forces the gradient components of the transmission image to have a long-tailed distribution and those of the reflection image to have a short-tailed distribution. Shih et al. [28] modeled the ghost effect using a double-impulse convolution kernel based on the characteristics of ghosting cues. They proposed an algorithm using the Gaussian mixture model (GMM) to separate the layers. Wan et al. [29] introduced a multi-scale depth of field (DoF) computation method and separated background and reflection using edge pixel classification based on the computed DoF map. Wan et al. [30] used content priors and gradient priors to automatically detect regions with reflections and those without reflections. They proposed an integrated optimization framework for content restoration, background-reflection separation, and missing content restoration.

2.2. Deep Learning (DL)

Specular Highlight Removal: Traditional non-DL methods removed reflection components using handcrafted priors. Nevertheless, one of the drawbacks of handcrafted priors is their lack of robustness in various environments. Recent research has shifted towards data-driven learning approaches to address this issue.
Funke et al. [31] proposed a GAN-based network to remove specular highlights from single endoscopic images. Lin et al. [32] introduced a new learning method: a fully convolutional neural network (CNN) for generating the diffuse component. Unlike traditional GAN methods, this approach used a multi-class classifier in the discriminator instead of a binary classifier to find more constrained features. Muhammad et al. [33] introduced Spec-Net and Spec-CGAN to remove high-intensity specular highlights in low-saturation images, particularly faces.
The DL techniques mentioned above can still result in color distortion in areas without specular components. Therefore, research has focused on detecting specular highlight regions and removing only the specular components in these areas. Fu et al. [34] proposed a generalized image formation model with region detection and a multi-task network based on it. Wu et al. [35] introduced a GAN network that models the mapping relationship between the two component image areas using an attention mechanism. Hu et al. [36] considered that specular highlight components have peculiar characteristics in the luminance channel and proposed a Mask-Guided Cycle-GAN. Wu et al. [37] presented an end-to-end deep learning framework consisting of a highlight detection network and a Unet-Transformer network for highlight removal. On the other hand, mask-guided techniques have limitations in removing large-area specular components.
Reflection Removal: Due to the success of deep convolutional neural networks in computer vision tasks compared to non-DL methods, researchers have proposed new data-driven methods for generating robust transmission image predictions for various reflection types. Fan et al. [38] introduced the first reflection removal deep learning model that utilizes linear methods to synthesize images with reflections for training and uses edge maps as auxiliary information for guidance. Zhang et al. [39] proposed an exclusion loss that minimizes the correlation between the gradient maps of the estimated transmission and reflection layers based on the observation that the edges of transmission and reflection images are unlikely to overlap. Yang et al. [40] introduced a deep neural network with a serial structure alternating between estimating reflection and transmission images. Li et al. [41] drew inspiration from iterative structure reduction and proposed a continuous network that iteratively refines the estimates of transmission and reflection images using an LSTM module. Zheng et al. [42] were inspired by the idea that the absorption effect can be represented numerically by the average of the refraction amplitude coefficient map. They proposed a two-step solution where they estimated the absorption effect from the intensity image and then used the intensity image and the absorption effect as inputs. Dong et al. [43] proposed a deep learning model to identify and remove strong reflection locations using multi-scale Laplacian features. Wei et al. [44] proposed method to enhance a baseline network with context encoding modules to use high-level context for clarity in reflective areas, and a novel loss function that leverages easier-to-collect misaligned real-world data. Li et al. [45] introduced a two-stage deep learning network that first estimates the reflection layer and then utilizes a Reflection-Aware Guidance module to improve the transmission layer prediction.

2.3. Training Datasets Construction

Specular Highlight Removal: As acquiring ideal specular and diffuse components in real-world settings is virtually impossible, constructing datasets has been carried out by constraining the shooting environment or applying constraints. Two datasets have been proposed for training deep learning models for specular highlight removal.
The first dataset, the specular highlight image quadruples (SHIQ) dataset [34], utilized the MIW dataset [46] acquired by capturing flashes from various directions, which was publicly available. They obtained reflection-free images from this dataset using existing multi-image-based reflection removal techniques [3]. They further selected high-quality images from the results. After unifying the specular regions to white, they constructed the dataset by cropping and producing specular region masks, resulting in intensity, diffuse, specular, and mask pairs. The second dataset, the PSD dataset [35], was constructed from real images using a physically based polarizing filter. They applied linear polarizing filters to the light source and circular polarizing filters to the camera, capturing images with minimal values in the filter regions. This process allowed them to obtain images with removed specular components on the object surface, forming the dataset.
Reflection Removal: Building a large-scale dataset containing reflection components is a challenging task. Therefore, various strategies have been used to produce training datasets, including generating mixed images by adding reflection and transmission images produced based on mathematical formulae. Alternatively, real-world reflection and transmission images have been acquired separately and combined through linear blending. Some approaches utilize real and synthetic images [38,47,48,49]. Datasets captured in various environments have been proposed to evaluate the network performance. Among them, the prominent dataset is S I R 2 [50]. S I R 2 is categorized into solid objects, postcards, and wild scenes. For solid objects and postcards, the images were captured from seven DoFs and three glass thicknesses in 20 scenes. Wild scenes comprised 100 scenes captured under various settings, including different glass thicknesses, camera settings, and uncontrolled illumination conditions.
Existing datasets for specular highlight removal each have their limitations as follows: The SHIQ dataset is built on the assumption that specular components are white, making it unable to remove specular components of various colors. The PSD dataset has low diversity due to its limited environments or objects and is composed of specular areas with narrow coverage, limiting the removal of specular areas of various sizes. Datasets for reflection removal are designed to eliminate reflection components, thus failing to consider specular components in the transmitted scene. For these reasons, this paper proposes the necessity of constructing a new form of dataset to eliminate integrated specular reflection for the first time.

3. Proposed Method

This section is structured to explain the methods for constructing the specular light (S-Light) dataset, designed to address the removal of specular components occurring in various scenarios, as well as to introduce the performance metric multi-scale normalized cross-correlation (MS-NCC) to measure the correlation between specular and diffuse components.
To address the limitations of existing datasets mentioned in the last paragraph of Section 2.3, we have constructed the database in two stages. Figure 4 illustrates an overall workflow for constructing our proposed S-Light DB. In the first step (step-1 in the figure), we collected 3D models that were rendered in Blender to deal with the specular highlight removal problem. Subsequently, in the next step (step-2 in the figure), we used methods [38,47] to handle the reflection removal issue.

3.1. Dataset Construction

As shown in Figure 5, the Case 1 dataset, S - Light R , consists of I, I t , and I r , whereas the Case 3 dataset, S - Light C , comprises I, I t d , and I r + I t s . Figure 6 shows example images for each case. Each subscript in S-Light represents surface specular, reflection, and combined, respectively.

3.1.1. Step 1: Construction of S - Light S Dataset

The SHIQ and PSD datasets consider Case 2 scenarios. Existing datasets have the following issues. The SHIQ dataset contains various objects but suffers from specular highlight colors composed only of white, making it less robust to various lighting conditions. PSD, being captured in a controlled laboratory environment, often exhibits mostly black backgrounds or repetitive patterns, and it lacks diversity in the types of objects.
Therefore, 130 multi-object 3D models were collected from the open-source platform Blend Swap [51]. The collected 3D models encompass a variety of object types with different materials (such as metals and plastics), textures (smooth and rough), and shapes (curved surfaces and sharp edges). These models were used in Blender (v3.4), a free and open-source 3D computer graphics software, using the Path Tracing-based Cycles engine [51], which employs materials and lighting grounded in physical laws, ensuring predictable and realistic outcomes. Seven hundred and nineteen scenes were generated in rendering scenes where objects and backgrounds do not overlap. Each scene was rendered into intensity and diffuse images at a resolution of 1024 × 1224 . For data augmentation, the images were cropped randomly to a 512 × 512 resolution or resized, resulting in a dataset of 3595 pairs.
Table 1 presents a quantitative comparison of the existing datasets, SHIQ, PSD, and the dataset developed in this study. The letters I, S, D, and M stand for intensity, specular, diffuse, and mask, respectively. “Clustering portion” indicates the proportion occupied by the largest cluster. Dimensionality reduction into two dimensions using t-SNE [52] and applying the DBSCAN clustering algorithm [53] was performed for each dataset. The t-SNE technique reduces dimensions in such a way that similar data points are clustered closely together, and DBSCAN clusters data points based on their density. Therefore, having many cluster groups with few points in each group suggests that the image dataset exhibits specific patterns and lacks diversity. Consequently, a higher “clustering portion” indicates that the dataset is more diverse and has a uniform density. Figure 7 provides examples of the results of applying the t-SNE and DBSCAN algorithms.
The distribution of the proportion of specular areas within the entire image for our constructed S - Light S , as well as for SHIQ and PSD, can be examined in Figure 8a. It demonstrates that, in contrast to S - Light S and SHIQ, which feature regions (specular area) of various sizes, PSD is concentrated in distributions of narrower areas. Additionally, Figure 8 b–d directly show that, whereas S - Light S and PSD contain specular components in various colors, SHIQ is limited to a single color (white) for its specular reflections. In summary, S - Light S is a dataset that encompasses diverse image patterns, specular colors, and specular regions of various sizes, effectively addressing the limitations observed in SHIQ and PSD.

3.1.2. Step 2: Construction of S - Light R and S - Light C Dataset

Previous research has proposed various methods to generate synthetic images and build the reflection-removal dataset. In this study, the techniques reported elsewhere [38,47] were applied to generate reflection images. Figure 4 shows the Step 2 process of constructing S - Light R and S - Light C as a block diagram. To build S - Light R and S - Light C , the I and D from S - Light S , constructed in Step 1, were considered as I T according to Equation (3). For generating reflection images, the Pascal VOC dataset [54] was regarded as I R according to Equation (4). Additionally, for constructing the S - Light R and S - Light C dataset, we assumed that the glass was made of transparent material.
Kim et al. [47] obtained depth images of the image behind the glass and the image reflected on the glass. They proposed a technique that maps the original image onto an object file converted from the depth image to a 3D mesh file for rendering. In this study, the images behind the glass were used for the proposed dataset, and ZoeDepth [55] was used to render the dataset for obtaining depth maps. Fan et al. [38] proposed a technique that applies a Gaussian filter to the reflection data because the reflection images are blurry. To apply the technique proposed by Fan et al., we generated various reflection images based on the transmission images acquired from Kim et al. Building on these two methods, we constructed an image set consisting of 5754 pairs, including I t , I t d , and I r images. In conclusion, the S-light DB is composed of three categories: S - Light R , S - Light S , and S - Light C , with each dataset containing 5754, 3595, and 5794 pairs, respectively.

3.2. Multi-Scale NCC

Many researchers have traditionally used performance metrics like the peak signal-to-noise ratio (PSNR) and the structural similarity index (SSIM) that consider the similarity with reference images to compare the performance of specular highlight and reflection removal deep learning models. Due to the limitation of not being able to acquire diffuse images in real-world settings, there is a problem with using PSNR and SSIM as performance metrics for the technique. This issue is the biggest problem when considering integrated specular removal.
The problem addressed in this study was separating or removing integrated specular images from the intensity image. Therefore, performance metrics that consider Cases 2 and 3 are needed. We paid attention to the relationship of the components considered in the two removal tasks. In specular highlight removal, diffuse and specular components occurring on the object surface are independent, and the independence of diffuse and specular components has been widely applied in non-learning methods [14,17]. For this reason, it can generally be said that the correlation between diffuse and specular components is low. In reflection removal, the transmission and reflection components are considered to originate from different scenes, which generally results in a low correlation between them. Techniques that leverage this characteristic exist. Therefore, we aimed to quantitatively evaluate the performance of removal techniques by considering the correlation between the diffuse component or the transmitted diffuse component and the specular reflection component in Cases 2 and 3.
Due to the limitations of using similarity metrics such as PSNR and SSIM, which require a reference image, we cannot apply it to real images. Thus, while traditional metrics are useful for their simplicity and general applicability, MS-NCC’s independence from reference images make it a more effective measure for evaluating specular reflection removal techniques. Furthermore, general correlation techniques typically consider only the overall correlation between variables. Such correlations cannot account for the local region correlations between specular and diffuse components that occur in opaque objects. The spatial information reflected in MS-NCC addresses this issue and has the advantage of allowing visualization of the correlation between the two components. Figure 9 provides example images showing correlation maps based on the window size, depicting the results of applying MS-NCC to the output and residual images of the deep learning model.
Equation (5) expresses the average map, M ( i , j ) f , of pixels within a window of odd size ( k × k ) in the image f, where ( i , j ) denotes the pixel coordinates, and the ranges for u and v are from ( k 1 ) / 2 to ( k 1 ) / 2 .
M ( i , j ) f = 1 k 2 u v f ( i + u , j + v )
Equation (6) presents a formula that indicates the variability in the input image. When the input images are the same, it represents the local variance, whereas it corresponds to the local covariance when the input images are different.
V ( i , j ) f = 1 k 2 u v f ( i + u , j + v ) · g ( i + u , j + v ) M ( i , j ) k · M ( i , j ) g
Equation (7) represents the local normalized cross-correlation (Local-NCC) between the input images, f and g, based on their local information.
L - N C C ( k ) f , g = m e a n V ( i , j ) f , g V ( i , j ) f , f · V ( i , j ) g , g
The window size, denoted as k n , is determined based on the smaller dimension (either height or width) of the input image, and this k n window size will be the same size as 2 ( n + 1 ) × 2 ( n + 1 ) of the smaller dimension. Multi scale-NCC was calculated as the weighted sum of local-NCC:
M S - N C C = n = 1 5 λ n · L - N C C ( k n ) f , g
In Equation (8), the weight value, λ n has been uniformly set. The MS-NCC equation mainly considers the global independence of the transmission and reflection images and the local independence between the diffuse and specular components within objects.
The range of correlation lies between 0 and 1. A range of less than 0.2 from 0 indicates a very weak or no correlation, from 0.2 to less than 0.4 indicates a weak correlation, and from 0.4 to less than 0.6 indicates a moderate correlation. Ranges beyond this signify a very high correlation. Therefore, an MS-NCC output value of less than 0.2 signifies a successful separation. Figure 10 visualizes the difference between varying correlation values according to the above mentioned threshold ranges. The bottom images of D and S (residual between input and model output) are input of MS-NCC. Although the RGB image of residual seems dark, it does not mean that the pixel value of the residual image is absolute zero (See S from Figure 10).

4. Experiments

There are no reports on a single-image-based deep learning network that simultaneously addresses the removal of specular images on object surfaces and images containing reflections on glass, including transmitted specular components. Therefore, experiments were conducted using models proposed for the existing problems of specular highlight removal and reflection removal. We have followed the same training strategy and hyperparameters as mentioned in the main papers of the models we have used. The models used in this experiment (also discussed in Section 2) are as follows:
-
Specular Highlight Removal Network: [35,36,37]
-
Reflection Removal Network: [41,42,44,45]

4.1. Training and Evaluation in Specular Highlight Removal Network

This experiment assessed the generalizability of specular highlight removal networks and their suitability for training on the dataset constructed in this study. The common characteristics of the specular highlight removal networks used in the experiment include detecting the removal region and removing the reflection component based on the detected area. Mask images indicating specular locations are required to train detection-based deep learning models. This study did not construct mask images separately for the Case 2 dataset. Therefore, the technique proposed in the SHIQ [34] paper was applied to produce binary mask images for this dataset. The training and test datasets were split at an 80:20 ratio for the experiment.
Table 2 and Table 3 present the performance of models trained on each dataset when tested on their respective test datasets. This experiment adopted PSNR, SSIM, NCC, and MS-NCC as the performance metrics. The output image I d and the specular image I s ( = I I d ) of the model were used to compute NCC and MS- NCC; both images were converted to grayscale images.
In general, the detection-based specular highlight removal networks used in this experiment showed higher performance on the previously proposed datasets, SHIQ [34] and PSD [35], than the dataset constructed in this paper. This trend aligns with expectations. Initially, the SHIQ [34] and PSD [35] datasets were constructed assuming that specular images were composed mostly of white highlights or produced in controlled experimental environments using single lighting conditions. As a result, the existing SHIQ [34] and PSD [35] datasets share common features that can aid deep learning models in specular region detection training.
On the other hand, the dataset constructed in this study consisted of various lighting conditions and objects, resulting in a wide range of specular components. Experiments confirmed that specular region detection training was not as effective on the proposed dataset. Figure 11 provides examples of models trained with the proposed dataset during testing.

4.2. Training and Evaluation in Reflection Removal Network

This experiment assessed the generalizability of reflection removal networks and their suitability for training with the dataset constructed in this study. Table 3 summarizes the performance of models trained on each dataset when tested on their respective test datasets, following a similar structure to that in Table 2. The models trained with the current dataset exhibited better performance across PSNR, SSIM, NCC, and MS-NCC metrics than previously conducted specular removal models.
Furthermore, the NCC and MS-NCC values for the SHIQ [34] and PSD [35] test datasets either outperformed or closely matched those trained on the same dataset. Figure 12 presents example images from the model trained on the present dataset when applied to SHIQ [34] test images. Several network output images exhibited a more natural appearance and effectively removed specular components occurring on smooth materials, such as glass and metal, compared to the ground truth images.

4.3. Training on the Constructed Dataset and Evaluation on Real Images

This experiment assessed the “trends” in whether the deep learning models trained on the proposed dataset show significant performance in real-world scenarios. An experimental model designed for reflection removal was used to train the dataset proposed in this paper, which considers reflection and surface specular components. Several reasons support this choice. First, experiment 4.1 showed that a dataset focused solely on surface specular components is not suitable for training a specular highlight removal model. As illustrated in Figure 8, unlike SHIQ and PSD, S - Light S ’s features areas of various sizes and distribution of colors are unsuitable for specular highlight removal models that are designed to focus on detecting specular regions, leading to ineffective learning, as evidenced by Table 2. Second, according to the observation from experiment 4.2, the reflection removal models also showed versatility when applied to the SHIQ [34] and PSD [35] test datasets. Unlike specular removal methods, reflection removal models mainly consider global features of the images. Thus, the mentioned rationale for experiment 4.1 also confirms the findings from experiment 4.2. Third, the reflection components constructed in this paper are present throughout the entire image, causing detection-based specular removal networks to be unsuitable. Three distinct settings were used in the training methodology for evaluating specular highlight removal and reflection removal tasks:
1.
Half Setting: In the first setting, 50% of the data were randomly selected from the S - Light R and S - Light C datasets.
2.
Combined Setting: The second setting exclusively used the S - Light C dataset.
3.
Pretrained Setting: The third setting used off-the-shelf models, provided by the authors to evaluate the specular and reflection removal tasks.
In each setting, well-known reflection removal benchmarks were used to evaluate the models on real-world images in Cases 2 and 3: the SIR-Wild and Solid object datasets. The SIR dataset consisted of intensity, transmission, and reflection images. The transmission and intensity images were used for Cases 2 and 3, respectively. NCC and MS-NCC were used as evaluation metrics to compare the correlation between the model output image and the residual image because there are no reference images in the considered scenarios, similar to experiments 4.1 and 4.2.
Table 4 presents the quantitative evaluations of the Wild and Solid Object datasets. Overall, the correlation between the model output image and the residual image was lowest for models trained in the “Half setting”. However, on the other side, the models trained under “Combined setting” performed decently good compared to the “Pretrained setting” where it was more biased towards the reflection removal task.
Similarly, Figure 13 illustrates the outputs from each setting trained on the model proposed by Li et al. [45]. The figure illustrates that the model, when trained in a “Half setting” condition, effectively eliminates reflection and specular highlight components from the input images. Conversely, in the “Pretrained setting” condition, it struggles to fully remove specular highlights. This challenge arises because the datasets employed for the pretrained models are predominantly aligned with reflection removal, leading to difficulties in generalizing the task of specular highlight elimination. In the “Combined setting”, the model achieves moderate performance on the provided input images.
To assess the generalization performance of the proposed DB in the context of specular highlight removal, we conducted experiments using reflection removal models trained under the “Half setting” condition, which had previously demonstrated superior performance. The PSD dataset, acquired through a polarizing camera, was employed as the evaluation dataset. Given the availability of precise GT images, we utilized PSNR and SSIM. Additionally, we considered MS-NCC and NCC as evaluation metrics. Furthermore, we applied a pretrained model, previously utilized, to the PSD. Table 5 presents the performance across each metric. A notable result is that models trained under the “Half setting” condition show superior results in MS-NCC and NCC, but the pretrained model performs better in terms of PSNR and SSIM. We visually confirmed that the output images from the pretrained model showed no significant differences from the input images.
Consequently, we also measured the PSNR and SSIM values with the input image as a reference, the results of which are enclosed in parentheses in Table 5. Overall, PSNR recorded approximately 30, and SSIM about 0.97. These results imply a minimal difference in pixel values between the intensity images and GT images (diffuse component), directly relating to our previous mention (in Section 3.1.1) that the dataset primarily consists of specular components with narrow area sizes. Additionally, the models trained under the “Half setting” generally recorded lower scores in traditional performance metrics, suggesting that training solely on the proposed DB might be limited in removing local specular reflection components. As reflection removal models have tendency to consider an entire region, they lack the ability to learn features from the local regions of the images. Figure 14 visually presents the output results of the half setting model in Cases 2 and 3.

5. Conclusions

Specular reflection removal was categorized into three main scenarios, and a dataset tailored for training single-image-based deep learning models was proposed to address these scenarios. Experiments that were divided into two categories were conducted to investigate suitable networks for the proposed S-Light dataset: one using an area detector-based specular highlight deep learning model and the other using a reflection removal deep learning model. The reflection removal model trained on the S-Lights dataset exhibited generalizability, as confirmed by the previously proposed metrics, such as SHIQ and PSD. Furthermore, a quantitative measure called MS-NCC, which relies on the correlation between the diffuse and specular components, was used to evaluate the network’s performance. This measure was effective when applied to real-world images from datasets, such as the S I R 2 Wildscene and Solid object. Therefore, for the effective removal of specular reflections, it is essential to construct an appropriate dataset.
In conclusion, three major research areas necessary for comprehensive specular removal were addressed. First, the establishment of training datasets was achieved. Second, the research into deep learning models was suitable for integrated specular removal. Finally, a correlation metric that utilizes the relationship between specular and diffuse components was developed for more detailed quantitative evaluation in real-world scenarios. However, in our present work, we consider the dataset’s size as a limitation due to the need of a larger dataset considering various environmental conditions that can help the deep learning models to be more robust towards the real life images and increase the generalization ability. Future studies will also focus on proposing novel deep learning models for unified specular reflection removal, as the present deep learning networks are not able to tackle the local region information, causing potential non-efficient removal.

Author Contributions

S.J. contributed to the idea, presentation, and method, as well as constructing the dataset and writing the manuscript. O.J. worked on managing the dataset, editing the manuscript materials, and contributing to the construction of dataset. C.B., M.K., T.L., Y.J., H.S., H.K. and S.D. helped construct the whole dataset in parts. S.K. contributed by reviewing the dataset and giving in-depth analysis of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea(NRF) grant funded by the Korea government(MSIT) (No. IRIS RS-2023-00219725). This research was supported by Basic Science Research Program through the National Research Foundation of Korea(NRF) funded by the Ministry of Education(IRIS RS-2023-00240109).

Institutional Review Board Statement

Not applicable. Ethical review and approval were not required for this study.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in https://drive.google.com/file/d/1Ntjq4KTprVVum8ykYp3HQ9PNVGHijZ4v/view?usp=sharing (accessed on 18 March 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bayoudh, K.; Knani, R.; Hamdaoui, F.; Mtibaa, A. A survey on deep multimodal learning for computer vision: Advances, trends, applications, and datasets. Vis. Comput. 2021, 38, 2939–2970. [Google Scholar] [CrossRef] [PubMed]
  2. Summaira, J.; Li, X.; Shoib, A.M.; Li, S.; Abdul, J. Recent advances and trends in multimodal deep learning: A review. arXiv 2021, arXiv:2105.11087. [Google Scholar]
  3. Guo, X.; Cao, X.; Ma, Y. Robust separation of reflection from multiple images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2187–2194. [Google Scholar]
  4. Su, T.; Zhou, Y.; Yu, Y.; Du, S. Highlight Removal of Multi-View Facial Images. Sensors 2022, 22, 6656. [Google Scholar] [CrossRef] [PubMed]
  5. Xue, T.; Rubinstein, M.; Liu, C.; Freeman, W.T. A computational approach for obstruction-free photography. ACM Trans. Graph. 2015, 34, 79. [Google Scholar] [CrossRef]
  6. Li, Y.; Brown, M.S. Exploiting reflection change for automatic reflection removal. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2432–2439. [Google Scholar]
  7. Nayar, S.K.; Fang, X.S.; Boult, T. Separation of reflection components using color and polarization. Int. J. Comput. Vis. 1997, 21, 163–186. [Google Scholar] [CrossRef]
  8. Umeyama, S.; Godin, G. Separation of diffuse and specular components of surface reflection by use of polarization and statistical analysis of images. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 639–647. [Google Scholar] [CrossRef] [PubMed]
  9. Wen, S.; Zheng, Y.; Lu, F. Polarization guided specular reflection separation. IEEE Trans. Image Process. 2021, 30, 7280–7291. [Google Scholar] [CrossRef] [PubMed]
  10. Kong, N.; Tai, Y.W.; Shin, S.Y. High-quality reflection separation using polarized images. IEEE Trans. Image Process. 2011, 20, 3393–3405. [Google Scholar] [CrossRef] [PubMed]
  11. Kong, N.; Tai, Y.W.; Shin, J.S. A physically-based approach to reflection separation: From physical modeling to constrained optimization. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 209–221. [Google Scholar] [CrossRef]
  12. Agrawal, A.; Raskar, R.; Chellappa, R. Edge suppression by gradient field transformation using cross-projection tensors. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 2301–2308. [Google Scholar]
  13. Agrawal, A.; Raskar, R.; Nayar, S.K.; Li, Y. Removing photography artifacts using gradient projection and flash-exposure sampling. In Proceedings of the ACM SIGGRAPH 2005 Papers, Los Angeles, CA, USA, 31 July–4 August 2005; pp. 828–835. [Google Scholar]
  14. Shafer, S.A. Using color to separate reflection components. Color Res. Appl. 1985, 10, 210–218. [Google Scholar] [CrossRef]
  15. Klinker, G.J.; Shafer, S.A.; Kanade, T. The measurement of highlights in color images. Int. J. Comput. Vis. 1988, 2, 7–32. [Google Scholar] [CrossRef]
  16. Bajcsy, R.; Lee, S.W.; Leonardis, A. Detection of diffuse and specular interface reflections and inter-reflections by color image segmentation. Int. J. Comput. Vis. 1996, 17, 241–272. [Google Scholar] [CrossRef]
  17. Tan, R.T.; Ikeuchi, K. Separating Reflection Components of Textured Surfaces Using a Single Image. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 178–193. [Google Scholar] [CrossRef] [PubMed]
  18. Shen, L.; Machida, T.; Takemura, H. Efficient photometric stereo technique for three-dimensional surfaces with unknown BRDF. In Proceedings of the Fifth International Conference on 3-D Digital Imaging and Modeling (3DIM’05), Ottawa, ON, Canada, 13–16 June 2005; pp. 326–333. [Google Scholar]
  19. Yoon, K.J.; Choi, Y.; Kweon, I.S. Fast separation of reflection components using a specularity-invariant image representation. In Proceedings of the 2006 International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 973–976. [Google Scholar]
  20. Shen, H.L.; Cai, Q.Y. Simple and efficient method for specularity removal in an image. Appl. Opt. 2009, 48, 2711–2719. [Google Scholar] [CrossRef] [PubMed]
  21. Yang, Q.; Wang, S.; Ahuja, N. Real-time specular highlight removal using bilateral filtering. In Proceedings of the Computer Vision—ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Proceedings, Part IV 11. Springer: Berlin/Heidelberg, Germany, 2010; pp. 87–100. [Google Scholar]
  22. Kim, H.; Jin, H.; Hadap, S.; Kweon, I. Specular reflection separation using dark channel prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1460–1467. [Google Scholar]
  23. Fu, G.; Zhang, Q.; Song, C.; Lin, Q.; Xiao, C. Specular Highlight Removal for Real-world Images. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2019; Volume 38, pp. 253–263. [Google Scholar]
  24. Akashi, Y.; Okatani, T. Separation of reflection components by sparse non-negative matrix factorization. In Computer Vision—ACCV 2014: 12th Asian Conference on Computer Vision, Singapore, 1–5 November 2014, Revised Selected Papers, Part V 12; Springer: Berlin/Heidelberg, Germany, 2015; pp. 611–625. [Google Scholar]
  25. Levin, A.; Zomet, A.; Weiss, Y. Separating reflections from a single image using local features. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2004, Washington, DC, USA, 27 June–2 July 2004; Volume 1, p. I. [Google Scholar]
  26. Levin, A.; Weiss, Y. User assisted separation of reflections from a single image using a sparsity prior. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1647–1654. [Google Scholar] [CrossRef] [PubMed]
  27. Li, Y.; Brown, M.S. Single image layer separation using relative smoothness. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2752–2759. [Google Scholar]
  28. Shih, Y.; Krishnan, D.; Durand, F.; Freeman, W.T. Reflection removal using ghosting cues. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3193–3201. [Google Scholar]
  29. Wan, R.; Shi, B.; Hwee, T.A.; Kot, A.C. Depth of field guided reflection removal. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 21–25. [Google Scholar]
  30. Wan, R.; Shi, B.; Duan, L.Y.; Tan, A.H.; Gao, W.; Kot, A.C. Region-aware reflection removal with unified content and gradient priors. IEEE Trans. Image Process. 2018, 27, 2927–2941. [Google Scholar] [CrossRef] [PubMed]
  31. Funke, I.; Bodenstedt, S.; Riediger, C.; Weitz, J.; Speidel, S. Generative adversarial networks for specular highlight removal in endoscopic images. In Medical Imaging 2018: Image-Guided Procedures, Robotic Interventions, and Modeling; SPIE: Houston, TX, USA, 2018; Volume 10576, pp. 8–16. [Google Scholar]
  32. Lin, J.; El Amine Seddik, M.; Tamaazousti, M.; Tamaazousti, Y.; Bartoli, A. Deep multi-class adversarial specularity removal. In Proceedings of the Image Analysis: 21st Scandinavian Conference, SCIA 2019, Norrköping, Sweden, 11–13 June 2019; Proceedings 21. Springer: Berlin/Heidelberg, Germany, 2019; pp. 3–15. [Google Scholar]
  33. Muhammad, S.; Dailey, M.N.; Farooq, M.; Majeed, M.F.; Ekpanyapong, M. Spec-Net and Spec-CGAN: Deep learning models for specularity removal from faces. Image Vis. Comput. 2020, 93, 103823. [Google Scholar] [CrossRef]
  34. Fu, G.; Zhang, Q.; Zhu, L.; Li, P.; Xiao, C. A multi-task network for joint specular highlight detection and removal. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7752–7761. [Google Scholar]
  35. Wu, Z.; Zhuang, C.; Shi, J.; Guo, J.; Xiao, J.; Zhang, X.; Yan, D.M. Single-image specular highlight removal via real-world dataset construction. IEEE Trans. Multimed. 2021, 24, 3782–3793. [Google Scholar] [CrossRef]
  36. Hu, G.; Zheng, Y.; Yan, H.; Hua, G.; Yan, Y. Mask-guided cycle-GAN for specular highlight removal. Pattern Recognit. Lett. 2022, 161, 108–114. [Google Scholar] [CrossRef]
  37. Wu, Z.; Guo, J.; Zhuang, C.; Xiao, J.; Yan, D.M.; Zhang, X. Joint specular highlight detection and removal in single images via Unet-Transformer. Comput. Vis. Media 2023, 9, 141–154. [Google Scholar] [CrossRef]
  38. Fan, Q.; Yang, J.; Hua, G.; Chen, B.; Wipf, D. A generic deep architecture for single image reflection removal and image smoothing. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3238–3247. [Google Scholar]
  39. Zhang, X.; Ng, R.; Chen, Q. Single image reflection separation with perceptual losses. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4786–4794. [Google Scholar]
  40. Yang, J.; Gong, D.; Liu, L.; Shi, Q. Seeing deeply and bidirectionally: A deep learning approach for single image reflection removal. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 654–669. [Google Scholar]
  41. Li, C.; Yang, Y.; He, K.; Lin, S.; Hopcroft, J.E. Single image reflection removal through cascaded refinement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3565–3574. [Google Scholar]
  42. Zheng, Q.; Shi, B.; Chen, J.; Jiang, X.; Duan, L.Y.; Kot, A.C. Single image reflection removal with absorption effect. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13395–13404. [Google Scholar]
  43. Dong, Z.; Xu, K.; Yang, Y.; Bao, H.; Xu, W.; Lau, R.W. Location-aware single image reflection removal. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 5017–5026. [Google Scholar]
  44. Wei, K.; Yang, J.; Fu, Y.; David, W.; Huang, H. Single Image Reflection Removal Exploiting Misaligned Training Data and Network Enhancements. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  45. Li, Y.; Liu, M.; Yi, Y.; Li, Q.; Ren, D.; Zuo, W. Two-stage single image reflection removal with reflection-aware guidance. Appl. Intell. 2023, 53, 1–16. [Google Scholar] [CrossRef]
  46. Murmann, L.; Gharbi, M.; Aittala, M.; Durand, F. A dataset of multi-illumination images in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4080–4089. [Google Scholar]
  47. Kim, S.; Huo, Y.; Yoon, S.E. Single image reflection removal with physically-based training images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5164–5173. [Google Scholar]
  48. Ma, D.; Wan, R.; Shi, B.; Kot, A.C.; Duan, L.Y. Learning to jointly generate and separate reflections. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2444–2452. [Google Scholar]
  49. Wan, R.; Shi, B.; Li, H.; Duan, L.Y.; Tan, A.H.; Kot, A.C. CoRRN: Cooperative reflection removal network. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 2969–2982. [Google Scholar] [CrossRef] [PubMed]
  50. Wan, R.; Shi, B.; Duan, L.Y.; Tan, A.H.; Kot, A.C. Benchmarking single-image reflection removal algorithms. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3922–3930. [Google Scholar]
  51. Blend Swap. Available online: https://www.blendswap.com/ (accessed on 19 January 2023).
  52. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  53. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the KDD’96: Second International Conference on Knowledge Discovery and Data Mining, Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  54. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  55. Bhat, S.F.; Birkl, R.; Wofk, D.; Wonka, P.; Müller, M. ZoeDepth: Zero-shot Transfer by Combining Relative and Metric Depth. arXiv 2023, arXiv:2302.12288. [Google Scholar] [CrossRef]
Figure 1. Diffuse and specular reflection occurring on the surface of an object.
Figure 1. Diffuse and specular reflection occurring on the surface of an object.
Sensors 24 02286 g001
Figure 2. Diffuse and glass reflection.
Figure 2. Diffuse and glass reflection.
Sensors 24 02286 g002
Figure 3. Example images with specular reflection. The red box marks specular reflection on a transparent material, and the yellow box shows it on an opaque material: (a) case occurring in transparent and opaque materials, (b) case occurring in transparent materials, and (c) case occurring in opaque materials.
Figure 3. Example images with specular reflection. The red box marks specular reflection on a transparent material, and the yellow box shows it on an opaque material: (a) case occurring in transparent and opaque materials, (b) case occurring in transparent materials, and (c) case occurring in opaque materials.
Sensors 24 02286 g003
Figure 4. Overall workflow of S-Light DB construction. After each step, S - L i g h t S and S - L i g h t R & C datasets were produced. In the figure, we have used [47] for ’Method 1’ and [38] for ’Method 2’.
Figure 4. Overall workflow of S-Light DB construction. After each step, S - L i g h t S and S - L i g h t R & C datasets were produced. In the figure, we have used [47] for ’Method 1’ and [38] for ’Method 2’.
Sensors 24 02286 g004
Figure 5. (Left) Example image illustrating the captured reflection component from the camera. (Right) Cases 1 and 2 briefly represent the problems that reflection removal and specular highlight removal aim to address. Case 3 is an extension of Case 1, where the target is the diffuse component.
Figure 5. (Left) Example image illustrating the captured reflection component from the camera. (Right) Cases 1 and 2 briefly represent the problems that reflection removal and specular highlight removal aim to address. Case 3 is an extension of Case 1, where the target is the diffuse component.
Sensors 24 02286 g005
Figure 6. Example images from the proposed dataset.
Figure 6. Example images from the proposed dataset.
Sensors 24 02286 g006
Figure 7. Results of applying t−SNE and DBSCAN clustering: (a) PSD dataset, (b) SHIQ dataset, and (c) S - Light S dataset. The each color refers to the cluster ID from DBSCAN algorithm.
Figure 7. Results of applying t−SNE and DBSCAN clustering: (a) PSD dataset, (b) SHIQ dataset, and (c) S - Light S dataset. The each color refers to the cluster ID from DBSCAN algorithm.
Sensors 24 02286 g007
Figure 8. Visualizing the distribution of the specular component: (a) proportion of specular regions per image, and (bd) specular intensity across RGB channels, utilizing probability density functions (PDF).
Figure 8. Visualizing the distribution of the specular component: (a) proportion of specular regions per image, and (bd) specular intensity across RGB channels, utilizing probability density functions (PDF).
Sensors 24 02286 g008
Figure 9. In the z-direction, (a) model output image, (b) residual image, and (cf) correlation map after applying the small ( k 5 ) to large window ( k 2 ) respectively. Please refer to the L - N C C ( k n ) f , g in Equation (8).
Figure 9. In the z-direction, (a) model output image, (b) residual image, and (cf) correlation map after applying the small ( k 5 ) to large window ( k 2 ) respectively. Please refer to the L - N C C ( k n ) f , g in Equation (8).
Sensors 24 02286 g009
Figure 10. Comparative visualization of output images across correlation value ranges (Right: high correlation and Left: low correlation). The bottom pair of images are grayscale of diffuse (D) and specular (S).The bounding boxes (red and green) indicate specular reflection region.
Figure 10. Comparative visualization of output images across correlation value ranges (Right: high correlation and Left: low correlation). The bottom pair of images are grayscale of diffuse (D) and specular (S).The bounding boxes (red and green) indicate specular reflection region.
Sensors 24 02286 g010
Figure 11. Example images from the region detection-based model [35] M/M’ represents mask GT (ground truth) and mask output in each row.
Figure 11. Example images from the region detection-based model [35] M/M’ represents mask GT (ground truth) and mask output in each row.
Sensors 24 02286 g011
Figure 12. Examples of specular removal results [41,42,44,45] on the SHIQ test set after training the dataset. (GT: ground truth).Bounding boxes mean the region of specular reflection.
Figure 12. Examples of specular removal results [41,42,44,45] on the SHIQ test set after training the dataset. (GT: ground truth).Bounding boxes mean the region of specular reflection.
Sensors 24 02286 g012
Figure 13. Visualization of example outputs from each setting: half, pretrained and combined. Bounding boxes depict the regions of specular reflection.
Figure 13. Visualization of example outputs from each setting: half, pretrained and combined. Bounding boxes depict the regions of specular reflection.
Sensors 24 02286 g013
Figure 14. Half setting model results: (a) input and output images [41,42,44,45] of Case 3, (b) input and output images [41,42,44,45] of Case 2.
Figure 14. Half setting model results: (a) input and output images [41,42,44,45] of Case 3, (b) input and output images [41,42,44,45] of Case 2.
Sensors 24 02286 g014
Table 1. Seven hundred and nineteen scenes were generated in rendering scenes where the objects and backgrounds do not overlap. (I, S, D, and M stand for Intensity, Specular, Diffuse, and Mask).
Table 1. Seven hundred and nineteen scenes were generated in rendering scenes where the objects and backgrounds do not overlap. (I, S, D, and M stand for Intensity, Specular, Diffuse, and Mask).
Number of ImagesImage SetSpecular ColorClustering PortionSceneAcquisition
PSD [35]13,380I, DMultiple0.0072210Captured
SHIQ [34]10,825I, D, S, MSingle0.91016Captured + RPCA [3]
S-Lights3595I, DMultiple0.96719Rendering
Table 2. Evaluation of specular highlight removal models trained on each dataset when assessed on the corresponding test datasets. The training dataset is indicated next to the model names on the right. The test dataset is listed at the top. When the training and test datasets differ, they are highlighted in green. “Mean” represents the average value within the highlighted green area. The best-performing values for each model/test dataset combination are shown in bold.
Table 2. Evaluation of specular highlight removal models trained on each dataset when assessed on the corresponding test datasets. The training dataset is indicated next to the model names on the right. The test dataset is listed at the top. When the training and test datasets differ, they are highlighted in green. “Mean” represents the average value within the highlighted green area. The best-performing values for each model/test dataset combination are shown in bold.
S-Lights (Proposed DB)SHIQ [34]PSD [35]Mean
PSNR↑SSIM↑NCC↓MS-NCC↓PSNR↑SSIM↑NCC↓MS-NCC↓PSNR↑SSIM↑NCC↓MS-NCC↓PSNR↑SSIM↑NCC↓MS-NCC↓
Wu [37]S-Lights15.8950.6660.6140.46819.9540.8610.6430.72817.4170.7540.7310.87218.6850.8070.6870.800
SHIQ [34]22.6570.7850.5120.49830.7540.9500.2850.58123.9310.8880.4460.51623.2940.8370.4790.507
PSD [35]24.1430.8440.2350.32424.3740.9080.2490.23129.0490.9510.2010.21524.2580.8760.2420.277
Wu [35]S-Lights25.3720.8680.3090.41925.8170.9270.3210.45928.1130.9460.1110.11026.9650.9360.2160.284
SHIQ [34]24.5900.8440.3100.27132.5510.9610.1800.29225.9190.9300.2560.26225.2540.8870.2830.266
PSD [35]25.8700.8620.1500.16626.9710.9290.1750.16330.1550.9580.1370.13026.4200.8950.1620.165
Hu [36]S-Lights24.5750.8250.1960.28522.7320.8630.4070.45425.3980.8920.1890.22924.0650.8770.2980.342
SHIQ [34]21.4750.7760.3980.51727.2860.9440.1920.34524.3600.9010.1760.20822.9170.8390.2870.362
PSD [35]21.1020.7940.3510.49921.090.8620.4350.71924.6310.8900.1870.26721.090.8280.3930.609
Table 3. Evaluation of reflection removal models trained on each dataset when assessed on the corresponding test datasets. The training dataset is indicated next to the model names on the right. This table is identical in structure to Table 2.
Table 3. Evaluation of reflection removal models trained on each dataset when assessed on the corresponding test datasets. The training dataset is indicated next to the model names on the right. This table is identical in structure to Table 2.
S-Lights (Proposed DB)SHIQ [34]PSD [35]Mean
PSNR↑SSIM↑NCC↓MS-NCC↓PSNR↑SSIM↑NCC↓MS-NCC↓PSNR↑SSIM↑NCC↓MS-NCC↓PSNR↑SSIM↑NCC↓MS-NCC↓
Wei [44]S-Lights27.8690.8980.1840.22523.6210.8900.2240.24826.0300.9210.1000.20324.8260.9060.1620.225
SHIQ [34]23.5180.8100.5690.37232.2170.9600.4780.45222.8090.8860.4970.32423.1640.8480.5330.348
PSD [35]24.8240.8500.2110.28326.0110.9180.2410.28429.8150.9570.1780.14725.4180.8840.2260.284
Li [41]S-Lights16.6020.5280.3880.46412.6850.4900.5350.35013.8440.5600.5040.52013.2650.5250.5200.435
SHIQ [34]24.7610.8440.2090.45232.6340.9640.2000.41826.5000.9280.2480.53325.6300.8860.2290.492
PSD [35]24.3510.8530.1700.40524.2970.9190.2830.38227.5930.9490.0830.32224.3240.8850.2260.394
Zheng [42]S-Lights27.1620.8670.2660.24725.0540.8900.1750.24924.6190.8860.1180.21324.8370.8880.1470.231
SHIQ [34]24.1920.8050.3680.37235.3090.9650.1410.34224.7290.8970.3340.38624.4600.8510.3510.379
PSD [35]18.1770.7190.3550.34418.5860.7640.3260.51121.7620.9180.1970.17218.3820.7420.3400.427
Li [45]S-Lights25.7670.8500.2410.22925.3540.9260.1700.21324.9070.9260.1330.17825.1300.9260.1510.195
SHIQ [34]22.4910.7970.2310.17330.4910.9580.2100.23422.4400.9000.1750.29422.4650.8480.2030.234
PSD [35]23.7870.8330.2220.22723.3860.9100.2220.30928.8720.9570.1310.18823.5860.8720.2220.268
Table 4. Validation results in S I R 2 Wildscene and S I R 2 Solid object. M and T represent the Mix and Transmission image sets. Results with better performance in the half, combined and pretrained settings are indicated in bold.
Table 4. Validation results in S I R 2 Wildscene and S I R 2 Solid object. M and T represent the Mix and Transmission image sets. Results with better performance in the half, combined and pretrained settings are indicated in bold.
Model Wildscene M [50]Wildscene T [50]Solid Object M [50]Solid Object T [50]
NCC ↓MS-NCC ↓NCC ↓MS-NCC ↓NCC ↓MS-NCC ↓NCC ↓MS-NCC ↓
Wei [44]Half0.14370.24960.15470.21290.11810.17120.15880.1375
Combined0.16180.21530.14740.20290.26380.27440.24720.3444
Pretrained0.17090.29910.29470.38290.18860.24960.33170.2915
Li [41]Half0.13010.23760.14690.24510.12710.16100.15550.1254
Combined0.12770.17380.12250.16590.14480.16680.19910.2100
Pretrained0.37530.18860.42610.20510.47150.19370.51670.2231
Zheng [42]Half0.12840.20190.13820.20230.11610.12670.12290.1395
Combined0.14410.25890.14330.25880.14970.26500.16490.3547
Pretrained0.15050.25250.18240.31390.42940.38290.28930.3311
Li [45]Half0.16670.15490.16890.17860.13760.14480.08420.1354
Combined0.14840.20590.14380.25780.10710.20820.12290.1467
Pretrained0.16330.29510.20970.24590.24480.25880.31590.3054
Table 5. Validation results of PSD under half and pretrained settings. The values in parentheses represent the PSNR and SSIM when the input image is used as the reference.
Table 5. Validation results of PSD under half and pretrained settings. The values in parentheses represent the PSNR and SSIM when the input image is used as the reference.
Model PSD [35]
PSNR ↑SSIM ↑NCC ↓MS-NCC ↓
Wei [44]Half25.50000.91550.10370.2350
Pretrained27.9452 (29.8697)0.9452 (0.9740)0.20540.3957
Li [41]Half26.20590.92000.15200.2600
Pretrained25.6764 (26.6873)0.9423 (0.9619)0.38280.3459
Zheng [42]Half22.35270.86730.12960.2546
Pretrained26.4305 (30.6514)0.9407 (0.9760)0.27290.3356
Li [45]Half22.17380.90730.13950.2395
Pretrained26.6873 (30.1141)0.9389 (0.9695)0.47070.5057
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jo, S.; Jang, O.; Bhattacharyya, C.; Kim, M.; Lee, T.; Jang, Y.; Song, H.; Kwon, H.; Do, S.; Kim, S. S-LIGHT: Synthetic Dataset for the Separation of Diffuse and Specular Reflection Images. Sensors 2024, 24, 2286. https://doi.org/10.3390/s24072286

AMA Style

Jo S, Jang O, Bhattacharyya C, Kim M, Lee T, Jang Y, Song H, Kwon H, Do S, Kim S. S-LIGHT: Synthetic Dataset for the Separation of Diffuse and Specular Reflection Images. Sensors. 2024; 24(7):2286. https://doi.org/10.3390/s24072286

Chicago/Turabian Style

Jo, Sangho, Ohtae Jang, Chaitali Bhattacharyya, Minjun Kim, Taeseok Lee, Yewon Jang, Haekang Song, Hyukmin Kwon, Saebyeol Do, and Sungho Kim. 2024. "S-LIGHT: Synthetic Dataset for the Separation of Diffuse and Specular Reflection Images" Sensors 24, no. 7: 2286. https://doi.org/10.3390/s24072286

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop