This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Reflectance-Oriented Probabilistic Equalization
for Image Enhancement

Abstract

Despite recent advances in image enhancement, it remains difficult for existing approaches to adaptively improve the brightness and contrast for both low-light and normal-light images. To solve this problem, we propose a novel 2D histogram equalization approach. It assumes intensity occurrence and co-occurrence to be dependent on each other and derives the distribution of intensity occurrence (1D histogram) by marginalizing over the distribution of intensity co-occurrence (2D histogram). This scheme improves global contrast more effectively and reduces noise amplification. The 2D histogram is defined by incorporating the local pixel value differences in image reflectance into the density estimation to alleviate the adverse effects of dark lighting conditions. Over 500 images were used for evaluation, demonstrating the superiority of our approach over existing studies. It can sufficiently improve the brightness of low-light images while avoiding over-enhancement in normal-light images.

Index Terms —  2D histogram equalization, reflectance, Retinex model, contrast enhancement, image enhancement

1 Introduction

Image enhancement aims to enhance image contrast and reveal hidden image details. With the rapid development of digital imaging devices, the number of images out there and the demand for image enhancement has increased significantly. Commercial raster graphics editors require image editing expertise or considerable manual effort to produce satisfactory image enhancement. Therefore, it is essential to develop an automated image enhancement technique that adapts to different input lighting conditions. Existing approaches to image enhancement can be classified into model-based approaches and learning-based ones. We focus on model-based approaches, as they are more interpretable and do not need labeled training data.

In model-based approaches, histogram equalization (HE) has received the most attention. It derives an intensity mapping function such that the entropy of the distribution of output intensities is maximized. However, HE extends the contrast between intensities with large populations to a wider range, even if it is not semantically important. This issue has been addressed by incorporating spatial information into density estimation [1, 2, 3, 4, 5, 6, 7]. For example, 2DHE [2, 3] equalizes the 2D histogram of intensity co-occurrence so that the contrast between frequently co-occurring intensities is enhanced to a greater extent. CACHE [6] incorporates image gradients into histogram construction to avoid the excessive enhancement of trivial background. However, such spatial information is not discriminative enough, especially for low-light image areas. Their equalization schemes [2, 3, 6] also overemphasize the importance of the frequently co-occurring intensities, tending to cause precipitous brightness fluctuation in very dark or very bright image areas.

Another direction [8, 9, 10, 11, 12, 13, 14, 15, 16, 17] is based on the Retinex model. It takes an image as a combination of illumination and reflectance components, which capture global brightness and sharp image details, respectively. Some studies [8, 9, 10, 11, 12] assumed reflectance to be the desired enhancement output and obtained it by estimating and removing illumination. However, this strategy sometimes leads to excessively enhanced brightness. In other studies, LIME [15] assumes that the gamma correction of the reflectance is the ideal form of low-light image enhancement. In NPE [13] and NPIE [16], the illumination is enhanced with HE and recombined with the reflectance to reconstruct the enhanced image. Ren et al. [17] found that the illumination can be leveraged as the exposure ratio of a camera response function (CRF), and proposed a novel CRF-based image enhancement approach called LECARM. These approaches are valid for discovering dark image details. However, it is not easy for them to find a solution optimized for both low-light and normal-light images; they tend to overly amplify the brightness and saturation of normal-light images.

Here, we propose a novel 2DHE approach known as reflectance-oriented probabilistic equalization (ROPE), which allows for adaptive regulation of global brightness. ROPE assumes intensity occurrence and co-occurrence to be dependent and derives the distribution of intensity occurrence (1D histogram) by marginalizing over the distribution of intensity co-occurrence (2D histogram). This scheme builds a novel bridge between 1D and 2D histograms. Compared to related approaches such as CVC [2] (2DHE) and CACHE [6, 7] (1DHE), ROPE provides more adequate contrast enhancement and less noise amplification. Inspired by RG-CACHE [7], we define a novel 2D histogram by incorporating the local pixel value differences in image reflectance into the density estimation to alleviate the adverse effects of dark lighting conditions. Experiments show that ROPE outperforms state-of-the-art image enhancement approaches from both qualitative and quantitative perspectives.

2 Proposed Approach

2.1 Preliminaries

Given a color image 𝐂in\mathbf{C}_{\mathrm{in}}, its grayscale image 𝐀in\mathbf{A}_{\mathrm{in}} is defined as the max of its RGB components and is equal to its value channel in HSV space [18]. Let 𝐀out\mathbf{A}_{\mathrm{out}} be the enhanced image of 𝐀in\mathbf{A}_{\mathrm{in}} in ROPE. Let \circ and \oslash denote element-wise manipulation and division, respectively. The final output 𝐂out\mathbf{C}_{\mathrm{out}} is computed as

𝐂out=(𝐂in𝐀in)𝐀out.\mathbf{C}_{\mathrm{out}}=(\mathbf{C}_{\mathrm{in}}\oslash\mathbf{A}_{\mathrm{in}})\circ\mathbf{A}_{\mathrm{out}}. (1)

Let 𝐀in={a(q)}\mathbf{A}_{\mathrm{in}}=\{a(q)\} and 𝐀out={a^(q)}\mathbf{A}_{\mathrm{out}}=\{\hat{a}(q)\}, where a(q)[0,K)a(q)\in[0,K) and a^(q)[0,K)\hat{a}(q)\in[0,K) are the intensities of the pixels qq in the input and output, respectively. KK is the total number of possible intensity values (typically 256). Our goal is to find an intensity mapping function TT of the form a^(q)=T(a(q))\hat{a}(q)=T(a(q)) to produce the enhanced image.

Let oko_{k} be the event of a(q)=ka(q)=k (occurrence), where a(q)𝐀ina(q)\in\mathbf{A}_{\mathrm{in}} and k[0,K)k\in[0,K) is an intensity value. Let p(ok)p(o_{k}) be its probability. The 1D histogram of 𝐀in\mathbf{A}_{\mathrm{in}} can then be expressed by {p(ok)}\{p(o_{k})\}. Let P(ok)P(o_{k}) be the cumulative distribution function of p(ok)p(o_{k}). In HE, the intensity mapping function T()T(\cdot) is given by T(k)=KP(ok)1T(k)=KP(o_{k})-1. The problem is how to properly define p(ok)p(o_{k}).

2.2 Modeling of 1D Histogram as Marginal Probability

In ROPE, we define p(ok)p(o_{k}) based on the 2D histogram of 𝐀in\mathbf{A}_{\mathrm{in}}. Let ci,jc_{i,j} be the event of a(q)=ia(q)=i and a(q)=ja(q^{\prime})=j (co-occurrence), where q𝒩(q)q^{\prime}\in\mathcal{N}(q) and 𝒩(q)\mathcal{N}(q) is the set of coordinates in the local window centered on the pixel qq. Let p(ci,j)p(c_{i,j}) be its probability. Then the 2D histogram can be written as {p(ci,j)}\{p(c_{i,j})\} with i,j[0,K)i,j\in[0,K) and i<ji<j. The construction of this 2D histogram is discussed in Section 2.3.

Refer to caption
(a) CVC [2]
Refer to caption
(b) CACHE [6, 7]
Refer to caption
(c) ROPE
Fig. 1: Toy examples showing how the 2D histogram value p(ci,j)p(c_{i,j}) (blue square) is embedded in the 1D histogram. While previous approaches overemphasize the frequently co-occurring intensities ii and jj, ROPE allows for more adequate and moderate contrast enhancement by distributing p(ci,j)p(c_{i,j}) over all k(i,j]k\in(i,j].

In CVC [2], given two intensities ii and jj, their 2D histogram value p(ci,j)p(c_{i,j}) is voted into the bin of the larger intensity jj and added to the 1D histogram value p(oj)p(o_{j}), as illustrated in Fig. 1a. Similarly, CACHE [6, 7] votes p(ci,j)p(c_{i,j}) into the bins of both intensities (Fig. 1b). These schemes overemphasize p(oi)p(o_{i}) and/or p(oj)p(o_{j}) and thus tend to cause precipitous brightness fluctuation in very dark or bright image areas as shown in Figs. 2b and 2e. Instead, we aim to find a proper method to distribute p(ci,j)p(c_{i,j}) over all k(i,j]k\in(i,j] for more adequate contrast enhancement (Fig. 1c as well as Figs. 2c and 2f).

Our thinking is as follows. Wu et al. [6, 7] have revealed that in HE, for any k(0,K)k\in(0,K), the degree of contrast enhancement (CE) between k1k-1 and kk is ultimately proportional to p(ok)p(o_{k}). Thus, for any i,j[0,K)i,j\in[0,K), the degree of CE between ii and jj is proportional to k=i+1jp(ok)\sum_{k=i+1}^{j}p(o_{k}). Meanwhile, related studies [2, 3] suggested that in 2DHE, p(ci,j)p(c_{i,j}) needs to be defined in such a way that it is positively correlated with the requirement of CE between ii and jj. These two insights lead to the inference that p(ok)p(o_{k}) should be modeled such that p(ci,j)k=i+1jp(ok)p(c_{i,j})\propto\sum_{k=i+1}^{j}p(o_{k}). This confirms our motivation above: p(ci,j)p(c_{i,j}) should not be delegated to p(i)p(i) and/or p(j)p(j) alone, but should be distributed over all k(i,j]k\in(i,j].

Refer to caption
(a) Input
Refer to caption
(b) RG-CACHE [7]
Refer to caption
(c) ROPE
Refer to caption
(d) Input
Refer to caption
(e) RG-CACHE [7]
Refer to caption
(f) ROPE
Fig. 2: Comparison of RG-CACHE [7] and ROPE.

Now, we would like to build a bridge between p(ok)p(o_{k}) and p(ci,j)p(c_{i,j}). We assume that oko_{k} and ci,jc_{i,j} are dependent on each other. If we consider oko_{k} to be a marginal event, its distribution p(ok)p(o_{k}) can be obtained by marginalizing over p(ci,j)p(c_{i,j}):

p(ok)=i=0K2j=i+1K1p(ok|ci,j)p(ci,j),p(o_{k})=\sum_{i=0}^{K-2}\sum_{j=i+1}^{K-1}p(o_{k}|c_{i,j})p(c_{i,j}), (2)

where p(ok|ci,j)p(o_{k}|c_{i,j}) is the conditional probability of oko_{k} given ci,jc_{i,j}. This formula implies that the 1D histogram value p(ok)p(o_{k}) can be modeled as a weighted average of all 2D histogram values p(ci,j)p(c_{i,j}). The conditional probabilities p(ok|ci,j)p(o_{k}|c_{i,j}) act as weights.

Recall that it is necessary to determine p(ok)p(o_{k}) in such a way that p(ci,j)k=i+1jp(ok)p(c_{i,j})\propto\sum_{k=i+1}^{j}p(o_{k}). Therefore, it is reasonable to assume that oko_{k} and ci,jc_{i,j} are dependent on each other, i.e., p(ok|ci,j)0p(o_{k}|c_{i,j})\neq 0, if and only if k(i,j]k\in(i,j]; otherwise, oko_{k} and ci,jc_{i,j} are mutually exclusive and p(ok|ci,j)=0p(o_{k}|c_{i,j})=0. In view of this, we introduce a significance factor sks_{k} for all k[0,K)k\in[0,K) and define the conditional probabilities by

p(ok|ci,j)={skk=i+1jskif k(i,j]0otherwise.p(o_{k}|c_{i,j})=\begin{cases}\displaystyle\frac{s_{k}}{\sum_{k^{\prime}=i+1}^{j}s_{k^{\prime}}}&\text{if $k\in(i,j]$}\\ 0&\text{otherwise}.\end{cases} (3)

The intensity value kk, which requires a greater contrast between k1k-1 and kk, should have a greater value of sks_{k}, and vice versa. However, we have no idea which intensity values are more important than others. In this study, we propose to determine sks_{k} through an iterative method. Let t[1,τ]t\in[1,\tau] be the index of the iteration and τ\tau be the maximum number of iterations. When t=1t=1, we initialize sk(1)=1/Ks^{(1)}_{k}=\nicefrac{{1}}{{K}} for all k[0,K)k\in[0,K) and compute p(1)(ok)p^{(1)}(o_{k}) using Eq. 2. For all t>1t>1, we update sk(t)s^{(t)}_{k} with sk(t)=p(t1)(ok)s^{(t)}_{k}=p^{(t-1)}(o_{k}) and recalculate p(t)(ok)p^{(t)}(o_{k}) using the updated significance factor. In this way, we can guarantee that oko_{k} with a larger probability p(ok)p(o_{k}) tends to have a larger significance factor and thus tends to receive more contribution from the 2D histogram values p(ci,j)p(c_{i,j}). Empirically, we found that two iterations are sufficient for ROPE to achieve satisfactory performance.

Fig. 2 compares RG-CACHE [7] and our approach. Using RG-CACHE, dark and bright intensity values fluctuated sharply, causing blurred contrast (Fig. 2b) and eerie artifacts (Fig. 2e). In comparison, ROPE achieved more high-quality and reliable enhancement.

2.3 Embedding Reflectance in 2D Histogram

Next, we describe how to construct the 2D histogram of the grayscale image 𝐀in\mathbf{A}_{\mathrm{in}}. In CVC [2], the histogram value p(ci,j)p(c_{i,j}) is determined as the co-occurrence frequency of the intensity value jj in the local window centered on the pixel of intensity ii, further weighted by |ij||i-j|. RG-CACHE [7] directly constructs a 1D histogram by incorporating the gradient of image reflectance into the density estimation. In this study, we borrow the idea of RG-CACHE to mitigate the negative effect of dark lighting conditions but embed the image reflectance into a 2D histogram instead of a 1D histogram.

Refer to caption
(a) Input 𝐀in\mathbf{A}_{\mathrm{in}}
Refer to caption
(b) p(ci,j)p(c_{i,j}) for CVC
Refer to caption
(c) p(ci,j)p(c_{i,j}) for ROPE
Refer to caption
(d) 1D histogram p(ok)p(o_{k})
Refer to caption
(e) Mapping function TT
Refer to caption
(f) Reflectance 𝐑\mathbf{R}
Refer to caption
(g) 𝐀out\mathbf{A}_{\mathrm{out}} for CVC
Refer to caption
(h) 𝐀out\mathbf{A}_{\mathrm{out}} for PE
Refer to caption
(i) 𝐀out\mathbf{A}_{\mathrm{out}} for ROPE
Fig. 3: Comparison of ROPE with CVC [2] and PE. LABEL:sub@subf:3-2dh-cvc and LABEL:sub@subf:3-2dh-rope: 2D histograms whose xx and yy-axes are intensities in [0,256)[0,256); histogram values are shown in color on a log scale. In LABEL:sub@subf:3-pe, PE means the 2D histogram of CVC was used for equalization.

Let 𝐈\mathbf{I} and 𝐑\mathbf{R} be the illumination and the reflectance components of 𝐀in\mathbf{A}_{\mathrm{in}}, respectively. We first use the relative total variation (RTV) approach [19, 15] as an edge-preserving filter [20, 21, 22, 23, 24, 25] to estimate 𝐈\mathbf{I}. We then consider a modified Retinex model as the formation of 𝐀in\mathbf{A}_{\mathrm{in}}: 𝐀in=𝐈e𝐑\mathbf{A}_{\mathrm{in}}=\mathbf{I}\circ e^{\mathbf{R}} (where ee is the Euler number). Thus, the definition of the reflectance becomes

𝐑=ln(𝐀in𝐈).\mathbf{R}=\ln(\mathbf{A}_{\mathrm{in}}\oslash\mathbf{I}). (4)

Eq. 4 calculates 𝐑\mathbf{R} in the logarithmic domain. It reveals much richer objectness information hidden in the dark areas because logarithmic scaling magnifies the difference between small quantities, as shown in Figs. 3a and 3f.

Let 𝐑={r(q)}\mathbf{R}=\{r(q)\}, where r(q)r(q) is the reflectance value of the pixel q𝐀inq\in\mathbf{A}_{\mathrm{in}}. We calculate the 2D histogram values by

p(ci,j)=q𝐀inq𝒩(q)|r(q)r(q)|δa(q),iδa(q),jq𝐀inq𝒩(q)|r(q)r(q)|,p(c_{i,j})=\frac{\sum_{q\in\mathbf{A}_{\mathrm{in}}}\sum_{q^{\prime}\in\mathcal{N}(q)}|r(q)-r(q^{\prime})|\delta_{a(q),i}\delta_{a(q^{\prime}),j}}{\sum_{q\in\mathbf{A}_{\mathrm{in}}}\sum_{q^{\prime}\in\mathcal{N}(q)}|r(q)-r(q^{\prime})|}, (5)

where 𝒩(q)\mathcal{N}(q) is a 7×77\times 7 window centered on qq according to CVC, and δ,\delta_{\cdot,\cdot} is the Kronecker delta. In this study, it is assumed that p(ci,j)=p(cj,i)p(c_{i,j})=p(c_{j,i}) and p(ci,i)=0p(c_{i,i})=0 for all i,j[0,K)i,j\in[0,K).

Recall that p(ci,j)p(c_{i,j}) needs to be defined in such a way that it is positively correlated with the requirement of CE between ii and jj (Section 2.2). Eq. 5 satisfies this condition exactly; the 2D histogram values capture the local pixel value differences in reflectance and are sensitive to the presence of meaningful objects hidden in the dark. As shown in Fig. 3f, most large differences in reflectance are present in the foreground objects, e.g., the plants in the center and on the right and the ColorChecker. These objects obviously require greater brightness and contrast (high CE requirement) than the background. Conversely, the reflectance values of the background, where contrast is of less importance (low CE requirement), are rather smooth. Once the 2D histogram is constructed, it is substituted into Eq. 2 to iteratively calculate the 1D histogram. The input image is then enhanced by HE as described in Section 2.1.

Table 1: Quantitative assessment. For the first four metrics, higher statistics indicate better quality, while LOE is the opposite. The best statistics per column are shown in bold.
Approach DE EME PD PCQI LOE
No Enhancement 7.17 15.7 27.9 1.00 0
LIME [15] 7.08 13.0 27.6 0.88 156.5
NPIE [16] 7.33 17.4 27.8 0.99 98.9
LECARM [17] 7.11 12.2 25.0 0.90 299.9
KIND [26] 7.02 11.7 22.5 0.87 155.4
CVC [2] 7.49 23.5 32.3 1.06 0
RG-CACHE [7] 7.64 26.0 37.6 1.02 0
\star ROPE 7.62 32.3 40.1 1.04 0

Fig. 3a shows an example of input image 𝐀in\mathbf{A}_{\mathrm{in}}. Figs. 3b3e show the 2D/1D histograms and the intensity mapping function TT obtained using CVC, PE, and ROPE (PE indicates our approach proposed in Section 2.2, but with the 2D histogram of CVC used). Compared to CVC, ROPE has more emphasis on darker pixels, especially for important objects, thanks to the incorporation of reflectance. As shown in Fig. 3g, CVC made no significant changes to the input, whereas PE improved the visibility of image detail, thereby demonstrating the effectiveness of Eq. 2. By harnessing the reflectance effectively, ROPE further boosted brightness and contrast for the most satisfying image enhancement.

3 Experiments

In Section 2.3, we compared ROPE to CVC [2] and RG-CACHE [7]. In this section, we mainly compare ROPE with four state-of-the-art approaches: LIME [15], NPIE [16], LECARM [17], and KIND [26]. The first three are based on the Retinex model; the last one is based on deep learning. All these approaches were evaluated using 578 images from four datasets: LIME [15], USC-SIPI [27], BSDS500 [28], and VONIKAKIS [29]. The size of the local window used in ROPE was set to 7×77\times 7, and the maximum number of iterations τ\tau was set to two.

3.1 Qualitative Assessment

Refer to caption
(a) Input
Refer to caption
(b) LIME [15]
Refer to caption
(c) NPIE [16]
Refer to caption
(d) LECARM [17]
Refer to caption
(e) KIND [26]
Refer to caption
(f) ROPE
Refer to caption
(g) Input
Refer to caption
(h) LIME [15]
Refer to caption
(i) NPIE [16]
Refer to caption
(j) LECARM [17]
Refer to caption
(k) KIND [26]
Refer to caption
(l) ROPE
Refer to caption
(m) Input
Refer to caption
(n) LIME [15]
Refer to caption
(o) NPIE [16]
Refer to caption
(p) LECARM [17]
Refer to caption
(q) KIND [26]
Refer to caption
(r) ROPE
Refer to caption
(s) Input
Refer to caption
(t) LIME [15]
Refer to caption
(u) NPIE [16]
Refer to caption
(v) LECARM [17]
Refer to caption
(w) KIND [26]
Refer to caption
(x) ROPE
Fig. 4: Comparison of ROPE with state-of-the-art image enhancement (see supplemental material [30] for more examples). Images are from LIME [15], USC-SIPI [27], BSDS500 [28], and VONIKAKIS [29].

The enhanced images obtained with the compared approaches are shown in Fig. 4. KIND uses CNNs for image enhancement, but in a broader sense, it is also based on the Retinex model and so has similar performance to LIME, NPIE, and LECARM. Two common disadvantages of the four approaches are 1) a tendency to over-amplify brightness (first row) and 2) excessive saturation if color distortion was previously hidden in the dark areas of the input (second row). In comparison, ROPE does not suffer from these problems. The previous approaches essentially excel at enhancing low-light images, but there are exceptions, as shown in the third row. In this example, ROPE provided the most pleasant brightness, especially for the flowers in the center. All of these examples demonstrate the much greater adaptability and consistency of ROPE; our approach is capable of improving brightness sufficiently for dark images while avoiding excessive enhancement for normal-light images.

3.2 Quantitative Assessment

We objectively evaluated the image enhancement approaches using five metrics: discrete entropy (DE), EME, PD, and PCQI for contrast enhancement and LOE for naturalness. DE measures the amount of information in an image. EME [31] measures the average local contrast in an image. PD [32] measures the average intensity difference of all pixel pairs in an image. PCQI [33] measures the distortions of contrast strength and structure between input and output. LOE [13] measures the difference in lightness order between the input and enhanced images. The lightness order means the relative order of the intensity values of two pixels. For DE, EME, PD, and PCQI, higher statistics indicate better quality, while LOE is the opposite.

Table 1 shows the statistics averaged over 500 test images of BSDS500. Let us first focus on the contrast enhancement metrics. Since CVC, RG-CACHE, and ROPE are based on HE, they could maximize the range of intensity values and so achieved the highest scores. ROPE showed excellent contrast improvement capabilities, taking first place in EME and PD and second place in DE and PCQI. In comparison, none of the four approaches compared in Section 3.1 showed good performance.

In terms of image naturalness, all the approaches based on HE obtained the best LOE scores. These approaches have an inherent monotonicity constraint on the intensity mapping function TT, so that the contrast enhancement does not change the order of intensity values in all pixels. In comparison, the approaches based on the Retinex model have poorer scores because they alter or eliminate the illumination component of the image, which results in a large variation in the lightness order.

Computational Complexity. Consider the processing of a grayscale image 𝐀in\mathbf{A}_{\mathrm{in}} with H×WH\times{W} pixels and KK possible intensity values. The complexity of the reflectance estimation based on RTV [19, 15] is 𝒪(HW)\mathcal{O}(HW), which is the same as in LIME. The complexity of the 2D histogram construction (Eq. 5) is 𝒪(w2HW)\mathcal{O}({w^{2}}HW), where w2=72w^{2}=7^{2} is the local window size. The 1D histogram construction (Eq. 2) requires complexity 𝒪(K(K21)/6)\mathcal{O}(K(K^{2}-1)/6). Although it appears to be computationally intensive, the processing time can be greatly reduced by exploiting convolutional operations. Applying the intensity mapping function TT to 𝐀in\mathbf{A}_{\mathrm{in}} finally takes a complexity of 𝒪(HW)\mathcal{O}(HW). Given a color image, the complexities are the same as those described above, since ROPE is only applied to its intensity channel (Section 2.1).

4 Conclusion

In this study, a novel image enhancement approach called ROPE is proposed. In this approach, an image is decomposed into illumination and reflectance components. The local pixel value differences in reflectance are embedded in a 2D histogram that captures the probability of intensity co-occurrence. ROPE derives a 1D histogram by marginalizing over the 2D histogram, assuming that intensity occurrence and co-occurrence are dependent on each other. Finally, an intensity mapping function is derived by HE for image enhancement. Evaluated on more than 500 images, ROPE surpassed state-of-the-art image enhancement approaches in both qualitative and quantitative terms. It was able to provide sufficient brightness improvement for low-light images while adaptively avoiding excessive enhancement for normal-light images.

References

  • [1] Tarik Arici, Salih Dikbas, and Yucel Altunbasak, “A histogram modification framework and its application for image contrast enhancement,” IEEE Trans. Image Processing, vol. 18, no. 9, pp. 1921–1935, 2009.
  • [2] Turgay Çelik and Tardi Tjahjadi, “Contextual and variational contrast enhancement,” IEEE Trans. Image Processing, vol. 20, no. 12, pp. 3431–3441, 2011.
  • [3] Turgay Çelik, “Two-dimensional histogram equalization and contrast enhancement,” Pattern Recognit., vol. 45, no. 10, pp. 3810–3824, 2012.
  • [4] Gabriel Eilertsen, Rafal K. Mantiuk, and Jonas Unger, “Real-time noise-aware tone mapping,” ACM Trans. Graph., vol. 34, no. 6, pp. 198:1–198:15, 2015.
  • [5] Haonan Su and Cheolkon Jung, “Low light image enhancement based on two-step noise suppression,” in ICASSP, 2017, pp. 1977–1981.
  • [6] Xiaomeng Wu, Xinhao Liu, Kaoru Hiramatsu, and Kunio Kashino, “Contrast accumulated histogram equalization for image enhancement,” in ICIP, 2017, pp. 3190–3194.
  • [7] Xiaomeng Wu, Takahito Kawanishi, and Kunio Kashino, “Reflectance-guided, contrast-accumulated histogram equalization,” in ICASSP, 2020, pp. 2498–2502.
  • [8] Daniel J. Jobson, Zia-ur Rahman, and Glenn A. Woodell, “Properties and performance of a center/surround Retinex,” IEEE Trans. Image Processing, vol. 6, no. 3, pp. 451–462, 1997.
  • [9] Daniel J. Jobson, Zia-ur Rahman, and Glenn A. Woodell, “A multiscale Retinex for bridging the gap between color images and the human observation of scenes,” IEEE Trans. Image Processing, vol. 6, no. 7, pp. 965–976, 1997.
  • [10] Giovanni Ramponi, “A cubic unsharp masking technique for contrast enhancement,” Signal Processing, vol. 67, no. 2, pp. 211–222, 1998.
  • [11] Frédo Durand and Julie Dorsey, “Fast bilateral filtering for the display of high-dynamic-range images,” ACM Trans. Graph., vol. 21, no. 3, pp. 257–266, 2002.
  • [12] Zeev Farbman, Raanan Fattal, Dani Lischinski, and Richard Szeliski, “Edge-preserving decompositions for multi-scale tone and detail manipulation,” ACM Trans. Graph., vol. 27, no. 3, 2008.
  • [13] Shuhang Wang, Jin Zheng, Hai-Miao Hu, and Bo Li, “Naturalness preserved enhancement algorithm for non-uniform illumination images,” IEEE Trans. Image Processing, vol. 22, no. 9, pp. 3538–3548, 2013.
  • [14] Xueyang Fu, Delu Zeng, Yue Huang, Xiao-Ping Zhang, and Xinghao Ding, “A weighted variational model for simultaneous reflectance and illumination estimation,” in CVPR, 2016, pp. 2782–2790.
  • [15] Xiaojie Guo, Yu Li, and Haibin Ling, “LIME: Low-light image enhancement via illumination map estimation,” IEEE Trans. Image Processing, vol. 26, no. 2, pp. 982–993, 2017.
  • [16] Shuhang Wang and Gang Luo, “Naturalness preserved image enhancement using a priori multi-layer lightness statistics,” IEEE Trans. Image Processing, vol. 27, no. 2, pp. 938–948, 2018.
  • [17] Yurui Ren, Zhenqiang Ying, Thomas H. Li, and Ge Li, “LECARM: Low-light image enhancement using the camera response model,” IEEE Trans. Circuits Syst. Video Techn., vol. 29, no. 4, pp. 968–981, 2019.
  • [18] Alvy Ray Smith, “Color gamut transform pairs,” in SIGGRAPH, 1978, pp. 12–19.
  • [19] Li Xu, Qiong Yan, Yang Xia, and Jiaya Jia, “Structure extraction from texture via relative total variation,” ACM Trans. Graph., vol. 31, no. 6, pp. 139:1–139:10, 2012.
  • [20] Li Xu, Cewu Lu, Yi Xu, and Jiaya Jia, “Image smoothing via L0{}_{\mbox{0}} gradient minimization,” ACM Trans. Graph., vol. 30, no. 6, pp. 174, 2011.
  • [21] Kaiming He, Jian Sun, and Xiaoou Tang, “Guided image filtering,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 35, no. 6, pp. 1397–1409, 2013.
  • [22] Qi Zhang, Xiaoyong Shen, Li Xu, and Jiaya Jia, “Rolling guidance filter,” in ECCV, 2014, pp. 815–830.
  • [23] Qi Zhang, Li Xu, and Jiaya Jia, “100+ times faster weighted median filter (WMF),” in CVPR, 2014, pp. 2830–2837.
  • [24] Takashi Shibata, Masayuki Tanaka, and Masatoshi Okutomi, “Gradient-domain image reconstruction framework with intensity-range and base-structure constraints,” in CVPR, 2016, pp. 2745–2753.
  • [25] Xiaoyong Shen, Chao Zhou, Li Xu, and Jiaya Jia, “Mutual-structure for joint filtering,” International Journal of Computer Vision, vol. 125, no. 1-3, pp. 19–33, 2017.
  • [26] Yonghua Zhang, Jiawan Zhang, and Xiaojie Guo, “Kindling the darkness: A practical low-light image enhancer,” in ACM Multimedia, 2019, pp. 1632–1640.
  • [27] sipi.usc.edu/database.
  • [28] Pablo Arbelaez, Michael Maire, Charless C. Fowlkes, and Jitendra Malik, “Contour detection and hierarchical image segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 33, no. 5, pp. 898–916, 2011.
  • [29] sites.google.com/site/vonikakis/datasets.
  • [30] www.kecl.ntt.co.jp/people/wu.xiaomeng/icassp2021/appendix.pdf.
  • [31] Sos S. Agaian, Blair Silver, and Karen A. Panetta, “Transform coefficient histogram-based image enhancement algorithms using contrast entropy,” IEEE Trans. Image Processing, vol. 16, no. 3, pp. 741–758, 2007.
  • [32] ZhiYu Chen, Besma R. Abidi, David L. Page, and Mongi A. Abidi, “Gray-level grouping (GLG): An automatic method for optimized image contrast enhancement - Part I: The basic method,” IEEE Trans. Image Processing, vol. 15, no. 8, pp. 2290–2302, 2006.
  • [33] Shiqi Wang, Kede Ma, Hojatollah Yeganeh, Zhou Wang, and Weisi Lin, “A patch-structure representation method for quality assessment of contrast changed images,” IEEE Signal Process. Lett., vol. 22, no. 12, pp. 2387–2390, 2015.