Next Article in Journal
Provably Secure Three-Factor-Based Mutual Authentication Scheme with PUF for Wireless Medical Sensor Networks
Previous Article in Journal
A Survey of Human Activity Recognition in Smart Homes Based on IoT Sensors Algorithms: Taxonomies, Challenges, and Opportunities with Deep Learning
Previous Article in Special Issue
Photographic Reproduction and Enhancement Using HVS-Based Modified Histogram Equalization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Photographic Reproduction Method Based on Feature Fusion and Virtual Combined Histogram Equalization

1
Graduate Institute of Automation Technology, National Taipei University of Technology, Taipei 106, Taiwan
2
Department of Computer Science and Information Engineering, National Taiwan University of Science and Technology, Taipei 106, Taiwan
3
Department of Electronic and Computer Engineering, National Taiwan University of Science and Technology, Taipei 106, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(18), 6038; https://doi.org/10.3390/s21186038
Submission received: 28 July 2021 / Revised: 6 September 2021 / Accepted: 8 September 2021 / Published: 9 September 2021
(This article belongs to the Special Issue Advanced Sensing for Intelligent Transport Systems and Smart Society)

Abstract

:
A desirable photographic reproduction method should have the ability to compress high-dynamic-range images to low-dynamic-range displays that faithfully preserve all visual information. However, during the compression process, most reproduction methods face challenges in striking a balance between maintaining global contrast and retaining majority of local details in a real-world scene. To address this problem, this study proposes a new photographic reproduction method that can smoothly take global and local features into account. First, a highlight/shadow region detection scheme is used to obtain prior information to generate a weight map. Second, a mutually hybrid histogram analysis is performed to extract global/local features in parallel. Third, we propose a feature fusion scheme to construct the virtual combined histogram, which is achieved by adaptively fusing global/local features through the use of Gaussian mixtures according to the weight map. Finally, the virtual combined histogram is used to formulate the pixel-wise mapping function. As both global and local features are simultaneously considered, the output image has a natural and visually pleasing appearance. The experimental results demonstrated the effectiveness of the proposed method and the superiority over other seven state-of-the-art methods.

1. Introduction

In the real world, the luminance intensity of environmental scenes has a very wide range. From glimmer starlight to blazing sunlight, the luminance variation could span over ten orders of magnitude. The human visual system (HVS) has an outstanding ability to adapt and perceive about 5~6 orders of magnitude. Previously, most consumer cameras can only capture nearly 2~3 orders of luminance variation. Due to the limited dynamic range, the captured image severely suffers from detail loss, especially in highlight and shadow regions [1].
With advancements in optical sensing, high dynamic range (HDR) sensors that can capture the entire luminance range of a real-world scene have been developed [2]. For example, some latest high-end digital single-lens reflex cameras, or some devised sensors including multiple sensor elements with different exposure levels, are able to capture entire details of both dark and bright parts of the scene simultaneously. Although the cost of such HDR sensors is high, the captured HDR images can contain a larger bit depth of the image than 24-bit depth per pixel. Typically, HDR images are stored in floating point and require 32-bit depth per pixel.
Despite the increasing availability of HDR images, presenting HDR scenes on traditional low dynamic range (LDR) or standard dynamic range (SDR) displays remains problematic because an SDR display can only display 256 brightness levels. The high cost and the under-developing display technology remain major obstacles to the mass production of HDR display devices. To solve this problem, the photographic reproduction method becomes an essential technique and has been a prominent subject in the field of image sensing research or the image-based applications [3].

2. Related Work

A photographic reproduction method should not only provide contrast adjustment but also the preservation of the luminance, details, and even the vividness of the colors of the original image. According to their modeling characteristics, most photographic reproduction methods are usually divided into the following three categories. First, the global-based photographic reproduction methods perform the same mapping function for all image pixels based on the global features of the input HDR image. In other words, an input pixel value produces a specific output value, regardless of its position. Both linear mapping functions and different nonlinear mapping functions are used to mimic the HVS. Reinhard and Devlin [4] proposed a global-based reproduction method using electrophysiology and the photoreceptor model that reflects human perception. It is a fast algorithm; however, the detail preserving is not considered. Mantiuk et al. [5] presented a piece-wise linear reproduction method that minimizes visible distortions by considering the penalty using the HVS contrast perception model. To reproduce optimal scene-referred images on a range of display devices, their method can adjust the image content by considering the display characteristics and surrounding illumination. However, the local operation, such as sharpening, is not considered. In [6], Kim et al. proposed integrating a modified weighted least square filter with mapping, which can preserve detail and maintain the global contrast through the competitive learning neural network. Furthermore, the color shift issue is solved by utilizing the Helmholtz–Kohlrausch effect in the light correction stage. Gommelet et al. [7] designed a global-based reproduction method to address the optimal rate-distortion problem, which typically occurs in the reconstruction of an HDR signal. In [7], a novel distortion model was built that takes the image gradient into account. Khan et al. [8] proposed a reproduction method that uses the features of HVS and the threshold versus intensity curve to adjust the individual bin-widths of the input image histogram. A global-based reproduction function is built using the modified histogram; however, the local information of an image is not used during the reproduction.
The overall pros of the global-based photographic reproduction methods are the capability to preserve the global contrast of the original images, and in addition, the computational complexity is low. However, the global compression of the dynamic range is typically accompanied by the suppression of local contrast, which is the inevitable con of the global-based methods. Moreover, especially for the scenes with a large difference in brightness, the bright and dark regions are the most severely sacrificed by detail loss because, compared to the entire dynamic range, the local intensity variation in such regions are almost ignorable.
To overcome the shortcomings of global-based methods, local-based photographic reproduction methods are proposed. For the local-based reproduction methods, different reproduction functions are designed to each pixel based on the pixel position and its surrounding pixels. Different pixel positions can share the same intensity but may relate to different reproduced values. Ahn et al. [9] proposed a local-based reproduction method, which utilizes center/surround retinex theory to adapt the local contrast. This work demonstrates superior contrast enhancement in the scenes with low log-average luminance. Nevertheless, in [9], sometimes over-enhancement occurs in the details. Tan et al. [10] proposed a halo-free reproduction method, which applies using an L0 smoothing filter to mimic the adaptability of the HVS mechanism. Cyriac et al. [11] presented a two-stage reproduction method, where the first stage is a simple gamma curve for global compression, and the second stage is a local-based reproduction scheme using psychophysical models for the visual cortex. However, the local-based processing in the second stage tends to degrade the overall naturalness. Croci et al. [12] proposed a reproduction method to reproduce the HDR video, where a tone-curve-space alternative is used as a substitute for the temporal per-pixel coherency to increase the computational efficiency. Li et al. [13] presented a clustering-based content and color adaptive reproduction method, which divides the input image into patches. By analyzing the local content information (e.g., patch mean, color variation, and color structure), the patches are formed into clusters, and the tone mapping is performed via a more compact domain. However, the patch-based processing tends to ignore the image information as a whole.
Due to considering local features, the typical pro of the local-based photographic reproduction methods is to provide more local details in dark and bright regions compared with global-based reproduction methods. Therefore, the local-based methods are suitable to the scenes where a large brightness difference exists. However, they are vulnerable to artifacts such as halo effects and block effects, which cause an unnatural overall appearance. Moreover, the global contrast decreases because of the lack of global features.
As both global- and local-based photographic reproduction methods have some drawbacks, some researchers have proposed decomposition-based photographic reproduction methods, which use decomposition techniques to obtain large-scale image structures (i.e., base layers) and small-scale image textures (i.e., detail layers), and thus, global-based (and local-based) approaches can be used for specific layers accordingly. Gu et al. [14] designed a local edge-preserving filter that has a locally adaptive property. Based on the filter, a retinex-based approach is presented, where the image is decomposed into one base layer and three detail layers for the reproduction manipulation. Barai et al. [15] proposed an HVS-inspired reproduction method, where the saliency map information is fed into the guided filter for image decomposition. Then, global compression and detail enhancement are performed in the base layer and the detail layer, respectively. Mezeni and Saranovac [16] presented an enhanced reproduction method, which decomposes the image into base/detail layers. Then, the base layer is scaled partially in the linear domain and partially in the logarithmic domain, and a detail enhancement is performed in the dark areas of the detail layer. However, when generating the output image, it is hard to fuse individual layers suitably. Liang et al. [17] presented a hybrid layer decomposition model for photographic reproduction, where a sparsity term is used to model the piecewise smoothness of the base layer, and the other sparsity term with a structural prior is used to model the piecewise constant effect of the detail layer. Miao et al. [18] presented a macro–micro-modeling reproduction method, in which multi-layer decomposition is utilized from the perspective of the micro model, and content-based global compression is utilized from the perspective of the macro model. The representative pro of the decomposition-based photographic reproduction methods is the flexibility to deal with different base and detail layers separately. However, the con of such a method is the difficulty of blending individual layers smoothly. That is, some blurs tend to occur in the final layer fusion process.
To exemplify the superiority of this work, Figure 1 shows a visual comparison among the global-based, local-based, decomposition-based methods, and our proposed method. In view of the abovementioned shortcomings of the global-based, local-based, and decomposition-based methods, this paper presents a new photographic reproduction method, which has the following three main advantages:
  • We propose using a hybrid histogram analysis scheme to extract mutually compatible global/local features in parallel, and a feature fusion scheme to construct the virtual combined histogram, which allows us to inherit the superiority of the global-based (and the local-based) methods smoothly.
  • Instead of performing late fusion (i.e., finally fusing all the processed layers as the decomposition-based methods do), the proposed virtual combined histogram equalization scheme can fuse global/local features in an earlier stage, which increases the naturalness of the output image.
  • Owning to the difference between the dark/bright regions and normal-luminance regions, we propose using the weight map to adaptively modify the weights locally in the feature fusion.

3. Proposed Approach

3.1. Pre-Processing for the Highlight/Shadow Detection

Figure 2 shows the overall framework of this study. The proposed method is designed due to the strategy of improving the visibility of highlight and shadow areas, while maintaining the global naturalness of the original image. In the pre-processing stage, a quick photographic reproduction method [19] is applied to the original HDR signal to obtain a pilot image ( I P i l o t ), which is a preliminary reproduced result with a simple global compression. Although the pilot image might suffer from detail loss locally, it is good enough for us to distinguish the dark/bright regions from the normal-luminance regions.
Subsequently, we modify the work of [20] for highlight/shadow detection as follows. First, a specular-free image ( I S F ) is defined as follows:
I c S F ( i , j ) = I c P i l o t ( i , j ) I D a r k ( i , j ) ,
where the subscript c R , G , B indicates one of the RGB color channels, and the dark channel ( I D a r k ) is defined as follows:
I D a r k ( i , j ) = min c R , G , B   I c P i l o t ( i , j )   .  
As I S F is obtained by subtracting the minimum of RGB values from I P i l o t , at least one of the three channels in I S F equals zero at each pixel position. Then, the modified specular-free (MSF) image is obtained by adding the average of the dark channel image to the specular-free image as follows:
I c M S F ( i , j ) = I ¯ D a r k + I c S F ( i , j ) ,
In [20], the difference between the MSF image and the pilot image can be used to detect the highlight regions in the image. With this feature, we find that if we multiply a correction parameter ( θ ) with the threshold and compare it with the pilot image, we can also detect shadow regions. Therefore, the proposed highlight/shadow detection scheme can be expressed as follows:
pixel   ϵ   { highlight , if   δ c ( i , j ) > t h r   for   all   c shadow ,                       if   I c P i l o t ( i , j ) < θ · t h r   for   all   c midtone , otherwise   .  
where δ c = I c P i l o t I c M S F , θ = 0.8 is an empirical value (from our experiments, 0.75 θ 0.85   would produce accurate detection result), and the threshold value ( t h r ) is obtained by applying the Otsu method in the pilot image. The Otsu method is an automatic way of creating binarization in image processing, and we find that it is suitable to determine the threshold in Equation (4). Figure 3 shows an example of the highlight/shadow detection results, which will be used as the estimate of the steering weight coefficients in the feature fusion stage (described later in Section 3.4).

3.2. Luminance Separation and Initial Logarithmically Normalization

As luminance information is mainly affected by the dynamic range, distinguishing luminance and chrominance from the original HDR signal is a common approach in photographic reproduction. In this study, luminance information was extracted by converting from RGB color space to CIE XYZ color space through the ITU-R BT.709 standard.
L i n = 0.2126 · I R H + 0.7152 · I G H + 0.0722 · I B H ,  
where I H indicates the input HDR signal and L i n indicates the corresponding luminance, which contains no chromatic information.
For different scenes, their dynamic range may vary quite greatly. To avoid the inconsistent dynamic range issue, the logarithmic function is a typical process to compress the luminance domain according to the following Weber–Fechner theory:
L log ( i , j ) = log 10 ( L i n ( i , j ) + 10 6 ) ,
where 10 6 is added to avoid the singularity error occurring as the input pixel luminance equals zero. Furthermore, to match the property that perceived brightness is proportional to the logarithm of the actual luminance intensity, its logarithmically normalized value can be expressed as follows:
L log _ n ( i , j ) = L log ( i , j ) min ( L log ( i , j ) ) max ( L log ( i , j ) ) min ( L log ( i , j ) )   ,  
where max ( L log ( i , j ) ) and min ( L log ( i , j ) ) represent the maximum and minimum values of L log ( i , j ) , respectively. To adapt various lighting conditions, the normalized logarithmic luminance value ( L log _ n ( i , j ) ), which always ranges between 0 and 1, are analyzed in the following steps.

3.3. Feature Extraction through Mutually Hybrid Histogram Analysis

The main challenge in photographic reproduction is to preserve both the global and local features of the original image, i.e., maintaining both the entire luminance balance and local detail information. In this study, the abovementioned features are neither the feature points used in computer vision, nor the feature vectors used in machine learning. The feature represents the general property of the entire image (i.e., global feature) or of individual local regions (i.e., local features) that are needed in the proposed photographic reproduction procedure.
Some proposed reproduction methods perform the global-based (and local-based) processes separately; in other words, they first apply a global luminance adaption and then perform local detail enhancement. However, we argue that this type of two-step strategy may not be the optimal solution because the goals of these two steps are inherently conflicting: one is to enhance the global features, and the other one is to enhance the local features.
As shown in Figure 4, we propose a parallel framework to simultaneously analyze the global histogram (constructed by the entire image) and local histogram (constructed by individual local image patches). The underlying concept of the proposed mutually hybrid histogram analysis is to extract the mutually compatible features from two statistical approaches.

3.3.1. Global Region Analysis and Global Feature Extraction

In global region analysis, the logarithmically normalized luminance plane is first transformed into a global histogram of K levels with equal bin width, where K is empirically set as one thousand. When divided by the total number of pixels in the image, the global histogram h G ( x k ) can be viewed as a probability density function of pixels. A parametric statistical method called the Gaussian mixture model (GMM) can then be used to structure h G ( x k ) as a weighted summation of three Gaussian functions as follows:
h G ( x k ) = n = 1 3 α n G · g ( x k , μ n G , σ n G ) ,
g ( x k , μ n G , σ n G ) = 1 σ n G 2 π exp [ ( x k μ n G ) 2 2 ( σ n G ) 2 ]   ,
where { x k , k = 0 , 1 , , K 1 } indicates the quantized reproduced levels of L log _ n , and α n G , n = 1 , 2 , 3 is the weight of the n -th Gaussian function. The reason for using three Gaussian functions to approximate h G ( x k ) is because in photographic reproduction, we normally concern the following three main parts: the highlight area, midtone area, and shadow area. From Equation (8), we refer to the global feature set as the following:
θ G = { α n G , μ n G , σ n G | n = 1 , 2 , 3 }     .  
The expectation-maximization (EM) algorithm [21] was adopted to solve the GMM estimation problem, which is used to find the maximum likelihood estimates of parameters in the statistical models involving unobserved latent variables. In this study, the likelihood function is defined as follows:
Likelihood ( θ G ) = ln [ K = 0 K 1 h G ( x k ) ] = K = 0 K 1 ln h G ( x k )   .  
To efficiently find the optimal θ G , the derivatives of the log-likelihood with respect to the initial α n G ,   μ n G , and σ n G are, respectively, set as zero (i.e., the expectation step), which yields a new parameter set of GMM (i.e., the maximization step). The EM algorithm iteratively switches between the expectation step and the maximization step until it converges (Please refer to [21] for the details of EM).

3.3.2. Local Region Analysis and Local Feature Extraction

In local region analysis, a sliding window scheme is adopted to visit each individual local region in raster scan order. Figure 5 illustrates the local region analysis, where a local region is of size M × M ( M = 129 as the default) and is centered at the current processing position ( i , j ) . Each local region is first divided into sixteen units with a size of 32 × 32 pixels, and 2 × 2 units constitute a partially overlapped subblock, e.g., the orange (or the green) square shown in Figure 5.
With consideration of the estimation accuracy and computation cost, each local region is subsampled into nine partially overlapping subblocks ( B n s u b ) that correspond to the two corner sets. First, the top-left (TL) corner set can be expressed by the following:
{ C n T L , n = 1 , 2 , , 9 }   .
where C 1 T L = ( i M / 2 ,   j M / 2 ) , C 2 T L = ( i M / 2 ,   j M / 4 ) , C 3 T L = ( i M / 2 ,   j + 1 ) , C 4 T L = ( i M / 4 ,   j M / 2 ) , C 5 T L = ( i M / 4 ,   j M / 4 ) , C 6 T L = ( i M / 4 ,   j + 1 ) , C 7 T L = ( i + 1 ,   j M / 2 ) , C 8 T L = ( i + 1 ,   j M / 4 ) , and C 9 T L = ( i + 1 ,   j + 1 ) . Second, the bottom-right (BR) corner set can be expressed by the following:
{ C n B R , n = 1 , 2 , , 9 }   .
where C 1 B R = ( i 1 ,   j 1 ) , C 2 B R = ( i 1 ,   j + M / 4 ) , C 3 B R = ( i 1 ,   j + M / 2 ) , C 4 B R = ( i + M / 4 ,   j 1 ) , C 5 B R = ( i + M / 4 ,   j + M / 4 ) , C 6 B R = ( i + M / 4 ,   j + M / 2 ) , C 7 B R = ( i + M / 2 ,   j 1 ) , C 8 B R = ( i + M / 2 ,   j + M / 4 ) , and C 9 B R = ( i + M / 2 ,   j + M / 2 ) . Each pair of ( C n T L , C n B R ) specifies the n -th subblock. To generate mutually compatible features (compatible to the global features) similar to Equation (8), this subsection aims to simulate each local histogram h L ( x k ) as a set of nine Gaussian functions g ( x k , μ n L , σ n L ) and to find the local feature set as the following:
θ L = { α n L , μ n L , σ n L | n = 1 , , 9 }   .  
Instead of using GMM, we adopt another statistical method called stratified sampling, in which the entire block is divided into homogeneous subblocks (defined as strata). The reason for partially overlapping is to avoid image artifacts such as the blocking effect and the halo effect. The distribution of each subblock is intentionally simulated as a Gaussian function, where the subblock mean and the subblock standard deviation are treated as the corresponding μ n L and σ n L in Equation (14), respectively. In addition, a spatial kernel (K) is used to weight the spatial correlation as follows:
K = α 1 L α 2 L α 3 L α 4 L α 5 L α 6 L α 7 L α 8 L α 9 L = 1 51 5 6 5 6 7 6 5 6 5   .
Inspired by [22], we adopted a summed-area table approach [23] to reduce the computation complexity of local region analysis as follows. First, the summed-area table ( T S A ) was generated by calculating the sum of all the pixels above and to the left of the current position as the following:
T S A ( i , j ) = i i , j j L log _ n ( i , j ) ,
Similar to Equation (16), the square summed-area table ( T S A 2 ) was generated by calculating the sum of all pixel squares as the following:
T S A 2 ( i , j ) = i i , j j L log _ n 2 ( i , j )
Notably, both T S A and T S A 2 could be efficiently computed through a one-pass procedure over the image by the following:
T S A p ( i , j ) = L log _ n p ( i , j ) + T S A p ( i , j 1 )                                                                             + T S A p ( i 1 , j ) T S A p ( i 1 , j 1 ) ,
where p = 1 and 2 .
Once the two summed-area tables were generated, the mean and standard deviation of each subblock could be quickly obtained by looking up T S A and T S A 2 because of the following closed-form solutions:
( Mean )   μ = 1 N T S A ( i 1 , j 1 ) + T S A ( i 0 1 , j 0 1 )                                                         T S A ( i 0 1 , j 1 ) T S A ( i 0 , j 1 1 ) ,
( Standard   Deviation )   σ = 1 N S μ 2 N ,
where N is the number of pixels in the subblock, and S = T S A 2 ( i 1 , j 1 ) + T S A 2 ( i 0 1 , j 0 1 ) T S A 2 ( i 0 1 , j 1 ) T S A 2 ( i 0 , j 1 1 ) . The four positions ( i 0 , j 0 ) ,   ( i 0 , j 1 ) ,   ( i 1 , j 0 ) , and ( i 1 , j 1 ) indicate the top-left, the top-right, the bottom-left, and the bottom-left corners of the subblock, respectively.

3.4. Virtual Combined Histogram Construction Based on Feature Fusion

Histogram equalization (HE) is a well-known method that by analyzing the histogram, pixel intensities can be arranged for enhancing the global contrast while maintaining image details by pursuing maximum entropy. As shown in the bottom row of Figure 4, both the global histogram and the highlight/shadow local histogram can be approximated (or characterized) as Gaussian mixtures. In this study, we propose a virtual combined histogram construction scheme based on nominally fusing the local/global Gaussian mixtures as follows.
First, considering that there is minor detail loss in the normal-luminance regions and more detail loss in the under-luminance (or over-luminance) regions during the reproduction process, the highlight/shadow detection result of Equation (4) is adopted to generate a binary map, where the highlight/shadow pixels are recorded as “1”, and the midtone pixels are recorded as “0”. A weight map function ( τ i , j ) is generated by convolving the binary map with a Gaussian low-pass filter (the Matlab inbuilt imgaussfilt function) to smooth the weighting difference. By doing so, we aim to make greater use of the local features (i.e., increase the weight map value in bright/dark regions) because the details of such regions are generally vulnerable to loss. The weight map function varies in different pixel positions because the weighting of local features should be region independent. Therefore, a virtual combined histogram is constructed by fusing global and local features through the following:
h i , j C o m b ( x k ) = ( ω 1 τ i , j ) · h G ( x k ) + ( ω 2 + τ i , j ) · h i , j L ( x k ) ,
where the subscript ( i , j ) indicate the pixel position, ω 1 and ω 2 represent the initial fusion weights (we set ω 1 = 0.4 and ω 2 = 0.6 empirically), and h G ( x k )   and h i , j L ( x k ) indicate the global and the local Gaussian mixtures, respectively. Moreover, we set an upper bound to constrain the maximum τ i , j value as 0.2. That is, the minimum weight to the global Gaussian mixtures in Equation (21) is guaranteed to be 0.2 to preserve the overall naturalness.

3.5. Luminance Modification and Color Recovery

Through the virtual combined histogram, a look-up table is generated in the traditional HE manner with linear interpolation. That is, the output luminance plane was modified by the following:
L o u t ( i , j ) = min ( L o u t ) + ( max ( L o u t ) min ( L o u t ) ) · CDF i , j ( x k ) ,
where L o u t is the adjusted luminance and CDF i , j ( x k ) is the Cumulative Distribution Function (CDF), which corresponds to the virtual combined histogram in Equation (21).
Overall, the pixel-wise modification function was controlled by manipulating both the global and local features through the virtual combined histogram. As each combined histogram was a summation of weighted Gaussian functions, the property of the Gauss error function was used to simplify the calculation by using the following:
Φ ( x k   |   μ , σ ) = 1 2 + 1 2 E r f ( x μ 2 σ ) ,
where Φ ( x k   |   μ , σ ) is the Gaussian CDF with parameters ( μ , σ ) . The beauty of the proposed virtual combined histogram scheme is that during the luminance modification process, only the global (and local) feature sets are used. Actually, the construction of an entire histogram is not needed.
Moreover, the Gauss error function E r f ( x ) can be approximated from [24] as the following:
tanh ( 77 x 75 + ( 116 25 ) tanh ( 147 x 73 ( 76 7 ) tanh ( 51 x 278 ) ) ) ,
where tanh is the hyperbolic tangent function. Finally, the output reproduced LDR image was obtained from restoring the color information by the following:
L D R R , G , B ( i , j ) = ( H D R R , G , B ( i , j ) L i n ( i , j ) ) s · L o u t ( i , j ) ,
where H D R R , G , B represents the three channel values of the original HDR image, L i n and L o u t represent the luminance before and after the proposed method, and s is the saturation factor (set as 0.6 in this study).

4. Experimental Results

In this section, we subjectively and objectively compare the effectiveness of the proposed method with those of the other photographic reproduction methods to confirm whether it affords more advantages than these methods. We selected seven classical and state-of-the-art methods for our experiments, including the following global-based reproduction method: Reinhard et al. [4] (published in 2005); the following three local-based reproduction methods: Ahn et al. [9] (published in 2013), Li et al. [13] (published in 2018), and Gao et al. [25] (published in 2020); and the following three decomposition-based reproduction methods: Gu et al. [14] (published in 2013), Liang et al. [17] (published in 2018), and Miao et al. [18] (published in 2019). For a comparison of the computational complexity, taking the image memorial (with size of 768 × 512) as an example, the processing time needed to generate a reproduced image is 0.252 s (in [4]), 0.533 s (in [9]), 5.301 s (in [13]), 0.627 s (in [25]), 0.788 s (in [14]), 2.189 s (in [17]), 0.733 s (in [18]), and 2.201 s (in the proposed method). All the methods are adjusted with the default parameters based on the suggestion of the original papers. In addition, the software and CPU are MATLAB R2016a and Intel Core i7, respectively.

4.1. Subjective Analysis

In subjective analysis, the performance of different methods can be judged through side-by-side visual comparison, such as according to the amount of regional detail information, the naturalness, etc. The simple baseline LDR images shown in Figure 6 indicate that a large luminance difference exists between the highlight and shadow areas of these test images; thus, many details are lost.
Figure 7 shows the reproduced results obtained using the Synagoguei test image. In Figure 7a, although the global brightness is balanced, the appearance of details is restricted by the global-based model. In Figure 7b, the regional scene performances in both the red and blue rectangles are poor and indistinct for the human eyes. In Figure 7c, the details of the shadow areas are preserved, whereas those of the bright area (such as the white dome) are almost imperceptible, and the tone of the entire image is monotonous and flat. In Figure 7d, the details of the red and blue rectangles are visible; however, the color of sky is oversaturated, resulting in a poor visual experience. In Figure 7e,f, although the details of the red and blue rectangles can be clearly seen, the naturalness is inevitably lost. As such methods are based on detail and base layer decomposition, image information tends to be overemphasized during decomposition and merging procedures. In Figure 7g, the details of the shadow areas (red and blue rectangles) are clear. However, the global contrast is unnatural: the highlight sky region is darkened, whereas the shadow areas are brightened, thus degrading the overall visual quality. In Figure 7h, our method shows advantages in preserving the details of the highlight and shadow areas because the proposed virtual combined histogram increases the pixel weights of local features for the highlight and shadow areas.
Figure 8 shows the reproduced results obtained using the Cadik_Desk02 test image. It is an indoor scene in which the lamp causes an extreme luminance difference in the captured image. In Figure 8a, the text on the book is barely perceptible because of the strong lighting. In Figure 8b,c, the detailed texture of the book is maintained; however, the global contrast in both figures is unbalanced, and the color tone is flat. In Figure 8d, the details are slightly preserved; however, some color shading occurs. In Figure 8e, the details are well retained; however, the overall appearance is over-sharpened (e.g., lampshade in red rectangle). This is because in the method of [14], the detail layer and base layer are processed separately, thereby overamplifying the detail information. In Figure 8f, the details are not evident (blue rectangle), and the global contrast is insufficient. In Figure 8g, although details are visible, the overall image appears unreal owing to the imbalance between the macro- and the micro-models. In Figure 8h, our method exhibits excellent naturalness of the image. Furthermore, because of the improved visibility of the highlight and shadow areas, more visual content is retained, and the overall color naturalness is satisfactory.
Figure 9 shows the reproduced results obtained using the C33_Store test image. In Figure 9a, the detailed information of the red and blue rectangles is lost as a result of global-based processing. In Figure 9b, although the details on the right side are more visible than those in Figure 9a, the regional details of the red rectangle are lost as a result of insufficient brightness. In Figure 9c,e, detailed information is perceptible but the degree of naturalness is low and the visual effects are not rich enough. In Figure 9e, enhanced smoothing is performed without consideration of the spatial correlation of the detail layer, leading to sharper and less-natural images. In Figure 9d, the detailed information of the red rectangle is slightly visible; however, the color is not vivid enough and lacks global contrast. In Figure 9f, although the overall appearance is natural, the visibility and sharpness in the red and blue rectangles areas are insufficient. In Figure 9g, the global contrast is good and the details of the highlight (i.e., blue rectangle) and shadow (i.e., red rectangle) areas are visible; nevertheless, the image is somewhat unnatural due to the lack of global contrast. In Figure 9h, our method demonstrates favorable visual richness because both the global and local characteristics are simultaneously considered through the construction of a virtual combined histogram. Consequently, the details in the highlight and shadow areas are clearly presented and the contrast and color naturalness of the entire image are improved.

4.2. Objective Analysis

In addition to the described subjective analysis, several objective quality indices were also applied to evaluate whether our method outperforms the other algorithms. The first quality index is called the tone-mapped image quality index (TMQI) [26]. The TMQI evaluates the quality of the reproduced images in terms of the following three aspects: structural similarity (TMQI-S), naturalness (TMQI-N), and overall quality (TMQI-Q) as follows. The TMQI-S value can be expressed by the following:
S = 2 σ x · σ y + C 1 σ x 2 + σ y 2 + C 1 · σ x y + C 2 σ x · σ y + C 2   ,
where σ x , σ y , and σ x y are the local standard deviations and the cross-correlation between the corresponding HDR and LDR patches; and C 1 and C 2 are the positive stabilizing constants. As suggested in [26], the local window size is set as 11 × 11 . The TMQI-N value can be expressed by the following:
N = P m P d / ρ ,
where ρ is a normalization factor, and P m and P d are the Gaussian and the Beta probability density functions, respectively. The TMQI-Q value can be expressed by the following:
Q = a · S α + ( 1 a ) · N β ,
where a is a weighting used to adjust the relative importance of the two terms ( a is set as 0.8011, as suggested in [26]); S and N indicate TMQI-S and TMQI-N values, respectively; and α and β indicate their sensitivities ( α = 0.3046 and β = 0.7088 , as suggested in [26]).
As shown in Equation (26), the TMQI-S is calculated using the standard deviations and cross-correlation between the HDR images and the reproduced results. As shown in Equation (27), the TMQI-N is calculated using Gaussian and Beta probability density functions that model the histograms of the means and standard deviations in the statistics conducted on massive natural images. As shown in Equation (28), the TMQI-Q is obtained from the weighted indices of structural similarity ( S value) and naturalness ( N value) by using a power function to adjust these two indicators. For the TMQI-S, the TMQI-N, and the TMQI-Q, a larger index value indicates a better quality of the reproduced result. Table 1 lists the results of comparisons using Figure 7, Figure 8 and Figure 9; apparently, the proposed method not only generates more visually pleasing reproduced results (as shown in Figure 7, Figure 8 and Figure 9), but also outperforms the other seven algorithms in terms of the average TMQI-S, TMQI-N, and TMQI-Q.
To further evaluate whether the proposed method is more effective than the other methods, we selected twenty-two test images from the datasets online [27,28,29,30]. Figure 10 shows some thumbnails of the test images, and Table 2 lists their names with the corresponding dynamic ranges. Moreover, four more objective quality metrics were added for conducting a thorough discussion. The first one was the feature similarity index for tone-mapped images (FSITM-TMQI) [31], an improved version of the TMQI that is based on a comparison of the phase-derived feature maps of the original HDR and the reproduced images. As in the case of the TMQI, a larger FSITM-TMQI value indicates a higher image quality. The second one was the dubbed blind/referenceless image spatial quality evaluator (BRISQUE) [32]. Unlike the TMQI and FSITM-TMQI, the BRISQUE is a no-reference quality assessment that evaluates the possible loss of naturalness in the spatial domain through scene statistics. The third one is the Blind TMQI (BTMQI) [33], another type of no-reference quality assessment that evaluates image quality by introducing features of statistical naturalness, structural preservation, and information entropy. For both the BRISQUE and BTMQI, lower values indicate less loss of overall naturalness, that is, better quality. The fourth one is the Integrated Local Natural Image Quality Evaluator (IL-NIQE) [34], which is a non-reference quality evaluation based on integrating multiple image statistics such as texture, color, and contrast. The IL-NIQE value reflects the global naturalness of the output image. The lower the IL-NIQE value is, the more natural it is.
The scatter plot in Figure 11 shows the detailed information of all twenty-two test images with each of the seven objective quality indices—TMQI-S, TMQI-N, TMQI-Q, FSITM-TMQI, BRISQUE, BTMQI, and IL-NIQE. This figure shows that the performance of the proposed method was among the top three for most evaluation indicators. Table 3 lists the average scores of the twenty-two test images obtained using different methods. With regard to the full-reference quality assessments (TMQI-S, TMQI-N, TMQI-Q, and FSITM-TMQI), our method obtained the best scores for these four assessments. The proposed method achieved the highest scores for the average TMQI-S, TMQI-N, and TMQI-Q, indicating that it achieved a strong balance between image structure and naturalness. In addition, our method also obtained the highest score for the average FSITM-TMQI, indicating that it generated more visually pleasing images based on the evaluation using phase-derived feature maps.
With regard to the no-reference quality assessments (BRISQUE, BTMQI, and IL-NIQE), our methods all obtained the best scores of the average BRISQUE, BTMQI, and IL-NIQE. By considering both global and local features to generate a virtual combined histogram, this method maintains the naturalness of an image and produces an output reproduced image with high image quality. Compared with global- and local-based reproduction methods that consider only global features (or only local features), our method can simultaneously take advantage of global and local features. Compared with the decomposition-based methods, our method does not need to process the base and detail layers separately, thus avoiding unnaturalness when blending different image layers. Overall, in Table 3, our method achieved the highest score in all seven assessments, indicating its excellent performance with natural-looking and rich information.
For the subjective analysis and evaluation, we invited 20 participants (10 males and 10 females) to take a subjective visual quality test. The participants were asked to rate the visual subjectiveness of all the images without knowing the applied method on the output images of twenty-two scenes using the eight comparative algorithms. The score ranges from 1 to 10 points, where 1 point means “unsatisfied” and 10 points means “excellent”. The mean and standard deviation of the mean opinion scores (MOS) of the subjective users are shown in Figure 12, where the proposed method is significantly better than the other methods.
In addition, the abovementioned FSITM-TMQI is actually obtained by averaging the scores of RGB channels, i.e., the FSITM-R, FSITM-G, and FSITM-B, respectively. The FSITM quality evaluation index is based on using the local phase similarity to construct a noise-independent feature map in the R, G, and B planes. In view of this, we further compare the average FSITM-R, FSITM-G, and FSITM-B using the twenty-two test images. As shown in Figure 13, our improved method performs better than the other seven reproduction methods in all the RGB channels of the FSITM, indicating that our method is not only really close to the real-world scene but also has an attractive visually pleasing character and natural color appearance.

5. Conclusions

Although HDR cameras are popularized in the digital photography industry, the current price of an HDR display is unaffordable to common people. Therefore, photographic reproduction techniques have great commercial potential due to the limited availability of HDR displays. This paper presented a new reproduction method, which considers global/local features simultaneously to achieve both global contrast-maintenance and local detail-preservation. Instead of performing the global-based and local-based processes separately, we combined two statistical approaches to extract the mutually compatible features to form a virtual combined histogram. In the feature fusion stage, a weight map is used to modify the importance between the global and local features. Moreover, with the integration of Gauss error function and global/local feature sets, the construction of an entire histogram is not actually needed in the luminance modification stage. From the experimental results, the proposed method outperforms other state-of-the-art methods in terms of various visual comparisons (Figure 7, Figure 8 and Figure 9) and objective evaluations (Table 1 and Table 3, Figure 11 and Figure 13). In the future, we plan to conduct the Wilcoxon test and the Friedman test to check whether the experimental results are statistically significant.

Author Contributions

Y.-H.L. and Y.-Y.C. carried out the studies and drafted the manuscript. K.-L.H. participated in its design and helped to draft the manuscript. I.-Y.C. and Y.-C.T. conducted the experiments and performed the statistical analysis. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was partially supported by the Ministry of Science and Technology, Taiwan, under Grant No. MOST 108-2221-E-027-095-MY2.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pan, Z.; Yu, M.; Jiang, G.; Xu, H.; Peng, Z.; Chen, F. Multi-exposure high dynamic range imaging with informative content enhanced network. Neurocomputing 2020, 386, 147–164. [Google Scholar] [CrossRef]
  2. Boukhayma, A.; Caizzone, A.; Enz, C. A CMOS image sensor pixel combining deep sub-electron noise with wide dynamic range. IEEE Electron Device Lett. 2020, 41, 880–883. [Google Scholar] [CrossRef]
  3. Garcia, D.V.; Rojo, L.F.; Aparicio, A.G.; Castello, L.P.; Garcia, O.R. Visual odometry through appearance- and feature-based method with omnidirectional images. J. Robot. 2012, 2012, 1–13. [Google Scholar] [CrossRef]
  4. Reinhard, E.; Devlin, K. Dynamic range reduction inspired by photoreceptor physiology. IEEE Trans. Vis. Comput. Graph. 2005, 11, 13–24. [Google Scholar] [CrossRef] [PubMed]
  5. Mantiuk, R.; Daly, S.; Kerofsky, L. Display adaptive tone mapping. ACM Trans. Graph. 2008, 27, 1–10. [Google Scholar] [CrossRef]
  6. Kim, B.; Park, R.; Chang, S. Tone mapping with contrast preservation and lightness correction in high dynamic range imaging. Signal Image Video Process. 2016, 10, 1425–1432. [Google Scholar] [CrossRef]
  7. Gommelet, D.; Roumy, A.; Guillemot, C.; Ropert, M.; Tanou, J.L. Gradient-based tone mapping for rate-distortion optimized backward-compatible high dynamic range compression. IEEE Trans. Image Process. 2017, 26, 5936–5949. [Google Scholar] [CrossRef]
  8. Khan, I.R.; Rahardja, S.; Khan, M.M.; Movania, M.M.; Abed, F. A tone-mapping technique based on histogram using a sensitivity model of the human visual system. IEEE Trans. Ind. Electron. 2018, 65, 3469–3479. [Google Scholar] [CrossRef]
  9. Ahn, H.; Keum, B.; Kim, D.; Lee, H.S. Adaptive local tone mapping based on retinex for high dynamic range images. In Proceedings of the 2013 IEEE International Conference on Consumer Electronics, Las Vegas, NV, USA, 11–14 January 2013; pp. 153–156. [Google Scholar]
  10. Tan, L.; Liu, X.; Xue, K. A retinex-based local tone mapping algorithm using L0 smoothing filter. In Proceedings of the Chinese Conference on Image and Graphics Technologies, Beijing, China, 19–20 June 2014; pp. 40–47. [Google Scholar]
  11. Cyriac, P.; Bertalmio, M.; Kane, D.; Corral, J.V. A tone mapping operator based on neural and psychophysical models of visual perception. In Proceedings of the Human Vision and Electronic Imaging XX, San Francisco, CA, USA, 17 March 2015; pp. 1–10. [Google Scholar]
  12. Croci, S.; Aydın, T.O.; Stefanoski, N.; Gross, M.; Smolic, A. Real-time temporally coherent local HDR tone mapping. In Proceedings of the 2016 IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016; pp. 1–5. [Google Scholar]
  13. Li, H.; Jia, X.; Zhang, L. Clustering based content and color adaptive tone mapping. Comput. Vis. Image Underst. 2018, 168, 37–49. [Google Scholar] [CrossRef]
  14. Gu, B.; Li, W.; Zhu, M.; Wang, M. Local edge-preserving multiscale decomposition for high dynamic range image tone mapping. IEEE Trans. Image Process. 2013, 22, 70–79. [Google Scholar] [CrossRef]
  15. Barai, N.R.; Kyan, M.; Androutsos, D. Human visual system inspired saliency guided edge preserving tone-mapping for high dynamic range imaging. In Proceedings of the 2017 IEEE International Conference on Image Processing, Beijing, China, 17–20 September 2017; pp. 1017–1021. [Google Scholar]
  16. Mezeni, E.; Saranovac, L.V. Enhanced local tone mapping for detail preserving reproduction of high dynamic range images. J. Vis. Commun. Image Represent. 2018, 53, 122–133. [Google Scholar] [CrossRef]
  17. Liang, Z.; Xu, J.; Zhang, D.; Cao, Z.; Zhang, L. A hybrid L1-L0 layer decomposition model for tone mapping. In Proceedings of the 2018 IEEE International Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1–9. [Google Scholar]
  18. Miao, D.; Zhu, Z.; Bai, Y.; Jiang, G.; Duan, Z. Novel tone mapping method via macro-micro modeling of human visual system. IEEE Access 2019, 7, 118359–118369. [Google Scholar] [CrossRef]
  19. Kim, M.; Kautz, J. Consistent tone reproduction. In Proceedings of the 10th IASTED International Conference on Computer Graphics and Imaging, Anaheim, CA, USA, 6 February 2008; pp. 152–159. [Google Scholar]
  20. Koirala, P.; Hauta-Kasari, M.; Parkkinen, J. Highlight removal from single image. In Proceedings of the International Conference on Advanced Concepts for Intelligent Vision Systems, Bordeaux, France, 28 September 2009; pp. 176–187. [Google Scholar]
  21. Lai, Y.; Chung, K.; Lin, G.; Chen, C. Gaussian mixture modeling of histograms for contrast enhancement. Expert Syst. Appl. 2012, 39, 6720–6728. [Google Scholar] [CrossRef]
  22. Liu, Y.-F.; Guo, J.-M.; Yu, J.-C. Contrast enhancement using stratified parametric-oriented histogram equalization. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 1171–1181. [Google Scholar] [CrossRef]
  23. Liu, Y.-F.; Guo, J.-M.; Lai, B.-S.; Lee, J.-D. High efficient contrast enhancement using parametric approximation. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 2444–2448. [Google Scholar]
  24. Vazquez-Leal, H.; Castaneda-Sheissa, R.; Filobello-Nino, U.; Sarmiento-Reyes, A.; Orea, J.S. High accurate simple approximation of normal distribution integral. Math. Probl. Eng. 2012, 2012, 1–22. [Google Scholar] [CrossRef] [Green Version]
  25. Gao, S.; Tan, M.; He, Z.; Li, Y. Tone mapping beyond the classical receptive field. IEEE Trans. Image Process. 2020, 29, 4174–4187. [Google Scholar] [CrossRef]
  26. Yeganeh, H.; Wang, Z. Objective quality assessment of tone mapped images. IEEE Trans. Image Process. 2013, 22, 657–667. [Google Scholar] [CrossRef] [PubMed]
  27. Anyhere Database. Available online: http://www.anyhere.com/ (accessed on 30 August 2020).
  28. Mignotte, M. Non-local pairwise energy-based model for the high-dynamic-range image compression problem. J. Electron. Imaging 2012, 21, 1–12. [Google Scholar] [CrossRef] [Green Version]
  29. Cadik Database. Available online: http://cadik.posvete.cz/tmo/ (accessed on 23 August 2020).
  30. Nemoto, H.; Korshunov, P.; Hanhart, P.; Ebrahimi, T. Visual attention in LDR and HDR images. In Proceedings of the 9th International Workshop on Video Processing and Quality Metrics for Consumer Electronics, Chandler, AZ, USA, 5–6 February 2015; pp. 1–6. [Google Scholar]
  31. Nafchi, H.Z.; Shahkolaei, A.; Moghaddam, R.F.; Cheriet, M. FSITM: A feature similarity index for tone-mapped images. IEEE Signal Process. Lett. 2015, 22, 1026–1029. [Google Scholar] [CrossRef] [Green Version]
  32. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  33. Gu, K.; Wang, S.; Zhai, G.; Ma, S.; Yang, X.; Lin, W.; Zhang, W.; Gao, W. Blind quality assessment of tone-mapped images via analysis of information, naturalness and structure. IEEE Trans. Multimed. 2016, 18, 432–443. [Google Scholar] [CrossRef]
  34. Zhang, L.; Zhang, L.; Bovik, A. A feature-enriched completely blind image quality evaluator. IEEE Trans. Image Process. 2015, 24, 2579–2591. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. A rough comparison among a global-based reproduction method [8] (bottom left), a local-based reproduction method [9] (top left), a decomposition-based reproduction method [18] (right), and our proposed method (middle). This example shows that our method inherits the advantages of both global- and local-based reproduction method, while avoiding the unnaturalness issue of decomposition-based reproduction method (due to processing different layers separately.
Figure 1. A rough comparison among a global-based reproduction method [8] (bottom left), a local-based reproduction method [9] (top left), a decomposition-based reproduction method [18] (right), and our proposed method (middle). This example shows that our method inherits the advantages of both global- and local-based reproduction method, while avoiding the unnaturalness issue of decomposition-based reproduction method (due to processing different layers separately.
Sensors 21 06038 g001
Figure 2. Flowchart of the proposed method, where the blue bold words indicate the section numbers.
Figure 2. Flowchart of the proposed method, where the blue bold words indicate the section numbers.
Sensors 21 06038 g002
Figure 3. An example of highlight/shadow detection. (a) The pilot image (pre-processed using the method of [19]). (b) Detection results from the proposed method, where the pink and cyan areas indicate highlights and shadows, respectively.
Figure 3. An example of highlight/shadow detection. (a) The pilot image (pre-processed using the method of [19]). (b) Detection results from the proposed method, where the pink and cyan areas indicate highlights and shadows, respectively.
Sensors 21 06038 g003
Figure 4. Proposed mutually hybrid histogram analysis approach. We analyze the global and local histograms simultaneously using different statistical methods, i.e., Gaussian mixture model for the former and stratified sampling for the latter. Although different statistical methods are applied, we aim to extract the mutually compatible features to form a virtual combined histogram (introduced later in Section 3.4).
Figure 4. Proposed mutually hybrid histogram analysis approach. We analyze the global and local histograms simultaneously using different statistical methods, i.e., Gaussian mixture model for the former and stratified sampling for the latter. Although different statistical methods are applied, we aim to extract the mutually compatible features to form a virtual combined histogram (introduced later in Section 3.4).
Sensors 21 06038 g004
Figure 5. Local region analysis, where the orange and the green squares, respectively, indicate the first and the second partially overlapped subblocks of a local region (the pink square with size of M × M ). At each processing position, local features are extracted by the statistical analysis of nine subblocks defined in Equations (12) and (13), and its simplification is achieved by utilizing the summed-area table defined in Equations (16) and (17).
Figure 5. Local region analysis, where the orange and the green squares, respectively, indicate the first and the second partially overlapped subblocks of a local region (the pink square with size of M × M ). At each processing position, local features are extracted by the statistical analysis of nine subblocks defined in Equations (12) and (13), and its simplification is achieved by utilizing the summed-area table defined in Equations (16) and (17).
Sensors 21 06038 g005
Figure 6. Baseline LDR images (processed using a simple linear compression), which illustrate the difficulty of photographic reproduction and can be compared with the results shown in Figure 7, Figure 8 and Figure 9. (a) Synagoguei. (b) Cadik_Desk02. (c) C33_Store.
Figure 6. Baseline LDR images (processed using a simple linear compression), which illustrate the difficulty of photographic reproduction and can be compared with the results shown in Figure 7, Figure 8 and Figure 9. (a) Synagoguei. (b) Cadik_Desk02. (c) C33_Store.
Sensors 21 06038 g006
Figure 7. Visual comparison using the test image Synagoguei. (a) Result of [4]. (b) Result of [9]. (c) Result of [13]. (d) Result of [25]. (e) Result of [14]. (f)Result of [17]. (g) Result of [18]. (h) Result of the proposed method.
Figure 7. Visual comparison using the test image Synagoguei. (a) Result of [4]. (b) Result of [9]. (c) Result of [13]. (d) Result of [25]. (e) Result of [14]. (f)Result of [17]. (g) Result of [18]. (h) Result of the proposed method.
Sensors 21 06038 g007aSensors 21 06038 g007b
Figure 8. Visual comparison using the test image Cadik_Desk02. (a) Result of [4]. (b) Result of [9]. (c) Result of [13]. (d) Result of [25]. (e) Result of [14]. (f) Result of [17]. (g) Result of [18]. (h) Result of the proposed method.
Figure 8. Visual comparison using the test image Cadik_Desk02. (a) Result of [4]. (b) Result of [9]. (c) Result of [13]. (d) Result of [25]. (e) Result of [14]. (f) Result of [17]. (g) Result of [18]. (h) Result of the proposed method.
Sensors 21 06038 g008aSensors 21 06038 g008b
Figure 9. Visual comparison using the test image C33_Store. (a) Result of [4]. (b) Result of [9]. (c) Result of [13]. (d) Result of [25]. (e) Result of [14]. (f)Result of [17]. (g) Result of [18]. (h) Result of the proposed method.
Figure 9. Visual comparison using the test image C33_Store. (a) Result of [4]. (b) Result of [9]. (c) Result of [13]. (d) Result of [25]. (e) Result of [14]. (f)Result of [17]. (g) Result of [18]. (h) Result of the proposed method.
Sensors 21 06038 g009
Figure 10. Thumbnails of partial test images with corresponding information provided in Table 2. First row from left to right: test images no. 1, no. 4, no. 6, and no. 5. Second row from left to right: test images no. 21, no. 17, and no. 3. Right side: test images no. 22. All the images are processed using the proposed method.
Figure 10. Thumbnails of partial test images with corresponding information provided in Table 2. First row from left to right: test images no. 1, no. 4, no. 6, and no. 5. Second row from left to right: test images no. 21, no. 17, and no. 3. Right side: test images no. 22. All the images are processed using the proposed method.
Sensors 21 06038 g010
Figure 11. Comparison of scatter plots using the twenty-two test images, where the horizontal axis indicates the image order, and the vertical axis indicates the objective quality index. (a) Result of TMQI-S. (b) Result of TMQI-N. (c) Result of TMQI-Q. (d) Result of FSITM-TMQI. (e) Result of BRISQUE. (f) Result of BTMQI. (g) Result of IL-NIQE.
Figure 11. Comparison of scatter plots using the twenty-two test images, where the horizontal axis indicates the image order, and the vertical axis indicates the objective quality index. (a) Result of TMQI-S. (b) Result of TMQI-N. (c) Result of TMQI-Q. (d) Result of FSITM-TMQI. (e) Result of BRISQUE. (f) Result of BTMQI. (g) Result of IL-NIQE.
Sensors 21 06038 g011aSensors 21 06038 g011b
Figure 12. Mean and standard deviation of subjective rankings of the eight comparative algorithms.
Figure 12. Mean and standard deviation of subjective rankings of the eight comparative algorithms.
Sensors 21 06038 g012
Figure 13. Comparison of the average FSITM-R, FSITM-G, and FSITM-B using the twenty-two test images.
Figure 13. Comparison of the average FSITM-R, FSITM-G, and FSITM-B using the twenty-two test images.
Sensors 21 06038 g013
Table 1. Comparison of TMQI-S, TMQI-N, and TMQI-Q using the test images shown in Figure 7, Figure 8 and Figure 9.
Table 1. Comparison of TMQI-S, TMQI-N, and TMQI-Q using the test images shown in Figure 7, Figure 8 and Figure 9.
TMQI-S (Structural Similarity)
Method[4][9][13][25][14][17][18]Ours
Synagoguei0.87240.84140.82540.71770.82660.79060.81910.9241
Cadik_Desk020.73380.74650.81160.78830.85160.78150.80370.9049
C33_Store0.93420.93260.89240.92350.89720.92550.90900.9072
Average0.84680.84020.84310.80980.85840.83260.84390.9121
TMQI-N (Naturalness)
Synagoguei0.50450.56900.28260.84920.31860.57850.79300.9113
Cadik_Desk020.13870.06150.82360.17900.85170.33490.70840.7269
C33_Store0.38090.60140.92390.85630.69990.95550.91040.9278
Average0.34140.41060.67670.62820.62340.62300.80390.8554
TMQI-Q (Overall Quality)
Synagoguei0.89100.89340.83690.90130.84440.88080.92260.9683
Cadik_Desk020.77810.76050.92510.80390.94030.83480.92510.9358
C33_Store0.88510.92300.96330.96010.92950.97500.96420.9663
Average0.85140.85900.90840.88840.90480.89690.93730.9568
Table 2. List of twenty-two test images and their dynamic ranges (D).
Table 2. List of twenty-two test images and their dynamic ranges (D).
No.NameDNo.NameD
1Belgium5.8712Cadik_Window5.10
2Fop_map4.1213C19_Casement2.46
3Mt. Tam West4.0614C21_Studio2.88
4Napa_Valley5.3615C22_Fort2.79
5Rend015.8416C29_Buildings3.52
6Still_Life3.9117C31_Parasol3.57
7Spheron_Siggraph5.0118C33_Store2.57
8Synagogue2.5819C37_Sculptures4.17
9Design Center5.2520C38_Cross3.65
10Cadik_Desk015.6821Spheron_PriceWestern3.73
11Cadik_Desk024.2622Memorial5.53
Table 3. Overall comparison of average TMQI, FSITM-TMQI, BRISQUE, BTMQI, and IL-NIQE using the twenty-two test images.
Table 3. Overall comparison of average TMQI, FSITM-TMQI, BRISQUE, BTMQI, and IL-NIQE using the twenty-two test images.
Method[4][9][13][25][14][17][18]Ours
TMQI-S0.81440.79460.80850.77370.81970.80660.81990.8606
TMQI-N0.36310.27650.63340.61430.58980.56890.72580.7805
TMQI-Q0.84640.81850.89060.87340.88380.87760.90460.9308
FSITM-TMQI0.83140.82650.84620.83400.84750.84870.85710.8784
BRISQUE28.989723.661628.282823.682722.916426.290518.947718.9413
BTMQI4.51534.09623.43664.38513.57763.61703.17372.7792
IL-NIQE27.069924.062625.644025.117622.092022.956122.803721.6186
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, Y.-H.; Hua, K.-L.; Chen, Y.-Y.; Chen, I.-Y.; Tsai, Y.-C. A New Photographic Reproduction Method Based on Feature Fusion and Virtual Combined Histogram Equalization. Sensors 2021, 21, 6038. https://doi.org/10.3390/s21186038

AMA Style

Lin Y-H, Hua K-L, Chen Y-Y, Chen I-Y, Tsai Y-C. A New Photographic Reproduction Method Based on Feature Fusion and Virtual Combined Histogram Equalization. Sensors. 2021; 21(18):6038. https://doi.org/10.3390/s21186038

Chicago/Turabian Style

Lin, Yu-Hsiu, Kai-Lung Hua, Yung-Yao Chen, I-Ying Chen, and Yun-Chen Tsai. 2021. "A New Photographic Reproduction Method Based on Feature Fusion and Virtual Combined Histogram Equalization" Sensors 21, no. 18: 6038. https://doi.org/10.3390/s21186038

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop