This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

SID-NISM: A Self-supervised Low-light Image Enhancement Framework

Lijun Zhang  Xiao Liu   Erik Learned-Miller   Hui Guan
University of Massachusetts Amherst
{lijunzhang, xiaoliu1990, elm, huiguan}@cs.umass.edu
Abstract

When capturing images in low-light conditions, the images often suffer from low visibility, which not only degrades the visual aesthetics of images, but also significantly degenerates the performance of many computer vision algorithms. In this paper, we propose a self-supervised low-light image enhancement framework (SID-NISM), which consists of two components, a Self-supervised Image Decomposition Network (SID-Net) and a Nonlinear Illumination Saturation Mapping function (NISM). As a self-supervised network, SID-Net could decompose the given low-light image into its reflectance, illumination and noise directly without any prior training or reference image, which distinguishes it from existing supervised-learning methods greatly. Then, the decomposed illumination map will be enhanced by NISM. Having the restored illumination map, the enhancement can be achieved accordingly. Experiments on several public challenging low-light image datasets reveal that the images enhanced by SID-NISM are more natural and have less unexpected artifacts.

1 Introduction

Generally, capturing high-quality images in dim light or back light conditions is challenging, since insufficient lighting can significantly degrade the visibility of images. Especially the lost details and low contrast would not only cause unpleasant subjective perceptions, but also hurt the performance of back-end computer vision systems which are designed for normal-light images. Though modern imaging sensors can automatically set high ISO, long exposure, and flash according to different circumstances to compensate for the low light, they suffer from different drawbacks. One solution is to use high dynamic range (HDR) imaging techniques, which has been integrated into modern cameras to tackle with dark light environment in the image acquisition stage. However, when it comes to restore existing poor-quality low-light images, HDR needs bunch of images under different illumination conditions as inputs, which limits its practical application.

Thus the low-light image enhancement has been a long-standing problem in the community with a great progress made over the past years. It can not only be used to increase the visual aesthetics of photos for people’s daily use, but also to stable the performance of many computer vision algorithms such as object detection. Existing methods have two mainstream ideologies. One tries to consider the low-light enhancement problem as an image decomposition problem based on the Retinex theory, in which each image can be separated into independent components and the image reflectance part can be seen as a reasonable enhancement result [4, 6, 17, 25]. However, these methods tend to introduce unexpected artifacts in the enhanced results. The other resorts to machine learning techniques based on large-scale image databases consisting of pairs of low-light image and corresponding enhanced image restored by image processing software like Photoshop [22, 3, 1, 18]. These methods have achieved impressive performance but usually limited by the quality and quantity of the training database.

To get rid of the restriction of using training data, in this paper, we present a novel self-supervised low-light image enhancement framework called SID-NISM, which could restore the quality of any single low-light image only relying on the visual information of the image itself. Our major contributions are summarized as follows.

(1) A self-supervised image decomposition network, SID-Net, is proposed to decompose the input image into lighting-independent reflectance, structure-aware smooth illumination and reflectance-related noise straightforwardly according to the robust Retinex theory [12]. Taking as inputs any given input image and its corresponding histogram equalization image, SID-Net can converge to the optimal decomposed maps within limited iterations, in which two novel loss terms related to the formation of the reflectance and noise maps together with several common-used basic terms are adopted to guide the decomposition procedure preciously. As pointed out above, SID-Net is a image-specific network without depending on any prior training or reference image, which greatly distinguish it from existing supervised-learning methods like Retinex-Net [22].

(2) A nonlinear illumination saturation mapping function (NISM) is constructed to refine the decomposed low-light illumination map. Specifically, it could brighten up the whole image to a proper lighting level on the premise of preserving the contrast between foreground and background, which combats the weaknesses of Gamma correction in contrast preservation and bright regions enhancement. The final normal-light result can be restored by combining the denoised reflectance with the enhanced illumination.

Comprehensive experiments are conducted to illustrate the performance of the proposed method. Human perception user study suggests that people are more inclined to prefer the output of our method and find less unexpected artifacts in our results when the results of multiple competing methods are presented in front of them, which is consistent with the objective evaluations. Both subjective and objective experiments demonstrate the superiority of our method over the state-of-the-art methods in producing natural and attractive enhancement results.

2 Related Work

Researchers have devoted their efforts to solving the problem of low-light image enhancement in the past decades. Many techniques have been developed to improve the quality of low-light images, which can be classified into two major categories as mentioned in Sect. 1: Retinex-based methods and learning-based methods.

Retinex-based Methods. According to the classic Retinex theory [10], an observed image can be decomposed into two components, reflectance and illumination, in which the former one can be taken as a kind of compelling low-light image enhancement result. Jobson et al. [7, 8] made some early attempts based on this model, but their results are usually unrealistic. Wang et al. [20] presented a naturalness preserved method by utilizing a lightness-order-error measure. However, it tends to produce results with dim artifacts and requires expensive computational cost. Fu et al. [4] and Ren et al. [17] succeed to simultaneously estimate reflectance and illumination maps through a weighted variational model and a sequential model respectively, but both suffered from preserving the property of Retinex model insufficiently. While Guo et al. [6] tired to only estimate the illumination map from the maximum values of three channels with constraints on preserving the main contour, then compute the reflectance by conducting element-wise division. And following the illumination estimation constraints in [6], Zhang et al. [25] proposed two more constraints for estimating illumination based on perceptually bidirectional similarity. These two methods have pretty good performance but usually introduce unexpected artifacts in the enhanced images. Besides, with the development of neural network, some researchers [26, 19, 22] sought to design image decomposition networks and illumination adjustment networks based on low-light and normal-light image pairs.

Learning-based Methods. Using machine learning tools to solve the problem of low-light image enhancement is a recent trend and also a promising direction. Dale et al. [3] first established a database comprising 1 million images. Given an input image to be enhanced, their system executes a visual search to find the closest images in the database; these images define the input’s visual context, which can be further exploited to instantiate the restoration operations. Kang et al. [9] constructed a database which stored the feature vectors describing training images along with vectors of enhancement parameters. Given a test image, the database was then searched for the best matching image, and the corresponding enhancement parameters were used to perform adjustment. Following the similar idea, Bychkovsky et al. [1] constructed a dataset containing 5,000 input-output image pairs that could be used to learn global tonal adjustments. To avoid collecting large-scale datasets, Shen et al. [18] trained their MSR-net designed upon the multi-scale Retinex theory on synthesized pairwise images.

With the power of neural network, these methods can get outstanding enhanced results, but at the same time it should be noticed that they are all based on supervised-learning frameworks and thus their performance highly depends on the quality and quantity of training datasets. Therefore Zhang et al. [24] proposed the first unsupervised-learning back-lit image restoration network based on the S-curve theory [23], which estimated the image specific S-curve through region-level optimal exposure evaluation. However, it is limited by the mid-gray assumption [6] in S-curve, which leads to unrealistic enhanced results with halos and gray shadows.

3 Retinex Theory

The classic Retinex theory [10] models the image compositions which could reflect the formation of low-light images to some extent. It assumes that the captured image can be decomposed into two components, reflectance and illumination. Let SS represents the source image, then the classic Retinex theory can be denoted by,

S=R×LS=R\times L (1)

where RR is reflectance, LL is illumination and ×\times is element-wise multiplication. Illumination LL refers to the various lightness on observed objects, while reflectance RR corresponds to the material RGB color that describes how objects reflect light, which is considered to be invariant to LL and other possible imaging conditions. In short, RR represents the intrinsic property of captured objects. Besides, in color images each channel of RR can be regarded as sharing the same grayscale illumination map LL.

Since low-light images usually suffer from darkness and unbalanced illumination distributions, the low-light image enhancement problem can be regarded as a procedure of estimating the illumination-independent reflectance according to the Retinex theory, which aims to remove the illumination effect and recover the original appearance of the scene objects. However, directly decomposing an input image into reflectance and illumination yields the intrinsic image decomposition problem [5], which is inherently ill-posed and may produce unrealistic results [6]. Therefore, some researchers [6, 4, 17, 22] converted the decomposition problem into an optimization problem and then computed the optimal solution through conventional optimization solvers or machine learning techniques. Moreover, instead of taking image reflectance as the enhancement result directly, existing methods usually project the enhanced illumination back to the reflectance by R×f(L)R\times f(L) at the end, where f()f(\cdot) stands for a manipulation operator adopted to enhance the illumination such as Gamma correction.

Another noticeable issue in low-light image enhancement is the unexpected noise raised by enhancing dark regions. Although the low-light source image doesn’t suffer from the noise problem, the noise hidden in the dark would be amplified along with stretching the contrast of dark regions. To address the intensive noise problem, researchers seek for help from denoising algorithms like BM3D [2] as the post-processing method, which is straightforward but not designed to solve the specific denoising problem here. Therefore we decide to follow the robust Retinex model [12], which integrate the noise term NN into the classic Retinex theory directly,

S=R×L+NS=R\times L+N (2)

Once the source image could be decomposed into the three components successfully, the reflectance map would be not only independent with the illumination map but also get rid of unexpected noise.

4 SID-NISM: Self-supervised Low-light Image Enhancement Framework

In this section, the proposed low-light image enhancement framework SID-NISM is presented in details. Fig. 1 illustrates the overall structure of SID-NISM, which consists of two stages, decomposition and enhancement, corresponding to the two main parts, SID-Net and NISM.

Refer to caption
Figure 1: The overall structure of SID-NISM. It consists of two stages, decomposition and enhancement, which are corresponding to the two main parts SID-Net and NISM.

4.1 SID-Net: A Self-supervised Image Decomposition Network

According to the definition of the Retinex theory, the reflectance component of an image should be invariant to the illumination component and other imaging conditions, which suggests that any pair of images with the same content should share the same reflectance regardless of their different illumination conditions. Thus inspired by the observation, a self-supervised learning image decomposition network called SID-Net is designed to split given low-light image into its reflectance, illumination and noise directly by taking as inputs the original image and its histogram equalization version. Notice that since the histogram equalization image is served as one of the paired images under different illumination condition to share the same reflectance, not the ground-truth normal-light images as in the supervised-learning methods, it doesn’t matter whether histogram equalization would produce images with unpleasant color distortion or unbalanced illumination. Besides, although theoretically capturing paired images or utilizing other image generation methods are acceptable as well, considering the practical application and the computational cost, histogram equalization image becomes the best option.

As illustrated in Fig. 1, SID-Net firstly takes as inputs the source image SlowS_{low} and its corresponding histogram equalization version SheS_{he} to split both of them into three components according to the robust Retinex theory, then iteratively converges to the optimal decomposition maps by minimizing the designed loss function. Since it is a self-supervised network without external information from training images, the key point of SID-Net is to devise reasonable and effective loss function to guide the network to generate expected decomposition maps. The loss function of SID-Net is formulated as follows,

=λrcrc+i=low,he(reci+λLLi+λRRi+λNNi)\mathcal{L}=\lambda_{rc}\mathcal{L}_{rc}+\sum_{i=low,he}{(\mathcal{L}_{rec}^{i}+\lambda_{L}\mathcal{L}_{L}^{i}+\lambda_{R}\mathcal{L}_{R}^{i}+\lambda_{N}\mathcal{L}_{N}^{i})} (3)

where λrc\lambda_{rc}, λL\lambda_{L}, λR\lambda_{R}, λN\lambda_{N} denote the coefficients to balance the reflectance similarities and the guidance of generating illumination, reflectance and noise maps. They are signi ficantly smaller than 1 to address the importance of the fidelity term rec\mathcal{L}_{rec} in the optimization, which are 0.01, 0.1, 0.001, 0.01 respectively in experiments. Notice that the reconstruction loss rec\mathcal{L}_{rec} and the three guidance loss terms L,R,N\mathcal{L}_{L},\mathcal{L}_{R},\mathcal{L}_{N} are both applied on the original low-light image SlowS_{low} and the histogram equalization image SheS_{he} to separate them into reflectance Rlow,RheR_{low},R_{he}, illumination Llow,LheL_{low},L_{he} and noise Nlow,NheN_{low},N_{he} in a self-supervised way.

Retinex reconstruction loss. First and foremost, according to the robust Retinex theory [12], the three decomposed maps should be able to reconstruct the original image. Thus the reconstruction loss rec\mathcal{L}_{rec} is formulated as,

rec=R×L+NS1\mathcal{L}_{rec}=\left\|R\times L+N-S\right\|_{1} (4)

It corresponds to the image formation, which could constrain the distance between estimated R×L+NR\times L+N and the source image SS.

Reflectance consistency loss. As described in Sect. 3, the reflectance reflects the intrinsic property of captured objects, which should be invariant to the scene lighting and imaging conditions. That is to say, in the case of the same captured objects, i.e. image content, the reflectance maps of the low-light image and its histogram equalization version should be as close as possible. Therefore, based on the above assumption, the reflectance consistency loss rc\mathcal{L}_{rc} is designed to constrain the reflectance similarities between RlowR_{low} and RheR_{he}.

rc=RlowRhe1\mathcal{L}_{rc}=\left\|R_{low}-R_{he}\right\|_{1} (5)

Illumination smoothness and consistency loss. The loss function designed for the illumination map L\mathcal{L}_{L} is consist of two parts, the smoothness term and the consistency term, which can be expressed as,

L=L×exp(αR)1+L×exp(αj=low,heLj)1\begin{split}\mathcal{L}_{L}&={\left\|\nabla L\times\exp(-\alpha\nabla R)\right\|_{1}}\\ &+\left\|\nabla L\times\exp(-\alpha\sum_{j=low,he}{\nabla L_{j}})\right\|_{1}\end{split} (6)

where \nabla denotes the gradient including h\nabla_{h} (horizontal) and v\nabla_{v} (vertical) and α\alpha denotes the coefficient balancing the strength of structure-awareness and edge-preservation which is set as 10 in experiments.

The first term is adopted to guide the illumination map to be piece-wise smooth in textural details while preserving the general structure of the original images. A common-used smoothness prior in various image restoration tasks, total variation minimization (TV) [14, 16], which minimizes the gradient of the whole image, is selected here as the fundamental of the illumination smoothness term. The second one is designed to preserve strong mutual edges while depressing weak ones. Similar to the illumination smoothness loss, it utilizes the weighted version of TV loss. Notice that since TV loss is structure-blindness, the original TV function is weighted with the gradient of the reflectance [22] and illumination maps [26] respectively to make the constructed loss be aware of the image structure and the two illumination maps be consistent with each other in terms of the image edges.

Reflectance contrast and color loss. To generate accurate reflectance map, in addition to the basic reconstruction loss, a novel reflectance contrast and color loss R\mathcal{L}_{R} is introduced to improve the image contrast and restore the image color as much as possible, which can be expressed as,

R=13ch=R,G,BRchβS^chF+RHSH2\begin{split}\mathcal{L}_{R}&=\frac{1}{3}\sum_{ch=R,G,B}{\left\|\nabla R^{ch}-\beta\nabla\hat{S}^{ch}\right\|_{F}}\\ &+\left\|R^{H}-S^{H}\right\|_{2}\end{split} (7)

where β\beta is the gradient amplification factor which is set as 10 in experiments. RHR^{H} and SHS^{H} are the hue channels of the reflectance map and the source image after converting them from the RGB to the HSV color space.

For the first term, as we know, low-light images always suffer from low contrast, which often indicates smaller gradient magnitudes [12]. Hence we attempt to manipulate the gradient magnitudes of the reflectance to boost the contrast of the enhanced results by amplifying the gradient of the input image with the factor β\beta. Notice that S^\nabla\hat{S} is a variant of the gradient of the input image S\nabla{S}, in which small gradients, i.e. the noise, are suppressed before amplification.

S^={0if |S|<ϵSotherwise`\nabla\hat{S}=\begin{cases}0&\text{if }|\nabla{S}|<\epsilon\\ \nabla{S}&\text{otherwise}\end{cases}` (8)

The second term is adopted to avoid color mismatch in the decomposed reflectance, which enforce the hue component of the reflectance map to be close enough to that of the source image.

Noise estimation loss. Although the reflectance contrast loss attempts to suppress noise by ignoring small gradients, noise may still appear in flat dark regions. It is necessary to guide the decomposition of the noise map by,

N=S×NF\mathcal{L}_{N}=\left\|S\times N\right\|_{F} (9)

which constrains the overall intensity of the noise [12].

As for the detailed network configuration, SID-Net first adopts a commonly used feature extractor, which could be any classic network such as ResNet, U-Net, or even a simple CNN structure. Then, three separate 3×33\times 3 convolutional layers project the image features into branches, namely reflectance RR, illumination LL, and noise NN respectively. At the end, sigmoid function is used to constrain both RR and LL in the range of [0, 1], while tanh function is adopted to simulate additive noise to make NN fall in [-1, 1].

4.2 NISM: Nonlinear Illumination Saturation Mapping Function

After obtaining the low-light illumination map LL of the given image SS by SID-Net, a manipulation operator would be adopted to enhance LL as described in Sect. 3. In the previous methods [6, 17, 25], researchers got used to refine the illumination map through Gamma correction, say L^=L1/γ\hat{L}=L^{1/\gamma}. When γ>1\gamma>1, the illumination map will be brightened up. In Fig. 2(a), the function curve of Gamma transformation with γ=2.2\gamma=2.2 is shown by the red line. It can be seen that pixels whose intensity are less than 20% of the maximum illumination value, namely 0.2 in the scale of [0, 1], will be brightened to 50%, while the pixels whose original intensity is large enough (>80%>80\%) will keep the themselves basically unchanged. This property of Gamma correction would definitely do favor to enhance the illumination map especially in those dark regions, but it also brings two major flaws: 1) the contrast of the original image is destroyed due to the over-brightening of dark areas; 2) the illumination level of the final enhanced image is still insufficient because the bright areas are barely changed. As shown in Fig. 2(c), the overall quality of the enhanced result is improved compared with the low-light input Fig. 2(b), but it haven’t been unable to reach the standard of ’normal-light’ in human perception.

Refer to caption

(a) Curves

Refer to caption

(b) Input

Refer to caption

(c) Gamma

Refer to caption

(d) NISM

Figure 2: Comparison between the Gamma correction and the proposed manipulation operator NISM. (a) are the two function curves when γ=2.2\gamma=2.2 and η=2.2\eta=2.2. (c) and (d) are the corresponding enhanced results when adopting Gamma correction and NISM on the illumination map of the low-light image (b) respectively.

To combat the existing issues of Gamma correction, we consider depressing the brightening of dark pixels while increasing the illumination saturation of bright pixels. More concretely, for dark pixels, it is still necessary to be brightened up but should be carried on at a relatively lower level as compared with Gamma correction. In other words, the slope of the illumination operator at pixels whose intensity are less than 0.2 should be smaller than that of Gamma correction. On the contrary, the slope at bright pixels should be larger than that of Gamma correction, which could increase the illumination levels of bright pixels so as to render the final enhanced results bright enough.

According to the above considerations, a novel manipulation operator called Nonlinear Illumination Saturation Mapping (NISM) is proposed, which is formulated as,

L^=1(1L)η\hat{L}=1-(1-L)^{\eta} (10)

where η\eta is the control parameter as γ\gamma in Gamma correction. The function curve of NISM with η=2.2\eta=2.2 is depicted in the blue line in Fig. 2(a). It can be seen that the curve shape of NISM is consistent with the previous analysis about the improved curve slope. Indeed, the constructed NISM and the Gamma correction are symmetrical about y=1xy=1-x.

The final enhanced image is generated by recombining the decomposed reflectance RR and the refined illumination L^\hat{L} by R×L^R\times\hat{L}. The corresponding result is shown in Fig. 2(d), which reflects that NISM could transform the decomposed illumination map of the original low-light image into a normal-light level.

In implementation η\eta is not fixed. It is computed by first clustering pixels of the low-light illumination map into two clusters, bright pixels and dark pixels, by KMeans. Then taking the minimum illumination value of bright pixels as threshold TT, η\eta is calculated by,

η=log(10.8)/log(1T)\eta=\log(1-0.8)/\log(1-T) (11)

which means to map the minimum illumination value of bright pixels to 0.8 under NISM.

5 Experimental Results

5.1 Experiment Settings

As a self-supervised framework, SID-NISM doesn’t need any training dataset or prior information. For any given low-light image, SID-Net could split it into reflectance, illumination and noise components directly within hundred iterations, then NISM would enhance the decomposed illumination to combine with the reflectance. Besides, all the coefficients mentioned in Sect. 4 would remain the same during the experiments, which means it is not necessary to adjust the parameters for different inputs.

Evaluations are performed on real-scene images from four public datasets, DICM [11], LIME [6], LOL [22], and MEF [13]. And the proposed method SID-NISM is compared with 6 state-of-the-art low-light image enhancement algorithms, including 1) a weighted variational model for simultaneous reflectance and illumination estimation (SRIE) [4], 2)illumination map estimation method (LIME) [6], 3) joint enhancement and denoising method via sequential decomposition (JED) [17], 4) perceptually bidirectional similarity based illumination estimation (PBS) [25], 5) deep learning based Retinex decomposition (Retinex-Net) [22], and 6) unsupervised-learning based back-lit image restoration network (ExCNet) [24].

In the following subsections, both subjective and objective experimental results will be exhibited. More experiments including the intermediate decomposition results and the analysis of the effect on the object detection task can be found in the supplementary materials.

5.2 Image Decomposition Results

Since Retinex-based low-light image enhancement methods consider the problem as an image decomposition problem, it is necessary to show our intermediate decomposition results firstly. Fig. 3 exhibits the decomposition results of two low-light images, Totoro and toy in LOL dataset. Their final enhanced results can be found in the main paper. In general, the proposed SID-Net can decompose reasonable reflectance, illumination and noise maps for given input images in a self-supervised way successfully. The decomposed reflectance maps contain sufficient color and structure details of the captured scene. The illumination maps reflect the lighting conditions as observed in human eyes. While the noise maps extract underlying noise component in dark regions.

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

(a) Source SS

Refer to caption

(b) Reflectance RR

Refer to caption

(c) Illumination LL

Refer to caption

(d) Noise NN

Figure 3: Intermediate decomposition results of two low-light images, Totoro and toy in LOL dataset.

Fig. 4 compares the decomposition results of the proposed image decomposition network SID-Net with three state-of-the-art Retinex-based methods, LIME, JED and Retinex-Net, on Bookshelf in LOL dataset and Balloon in MEF dataset.

It can be seen that all of the four methods can decompose reasonable reflectance and illumination maps for given input images as expected. However, there also exists some differences among the four methods. In the aspect of reflectance, it is obvious that LIME can restore more color and structure details of the original objects compared with the other methods, because the reflectance of LIME is calculated by conducting element-wise division on the source image and the estimated illumination, which preserve the Retinex theory to a large extent. As for the illumination maps, SID-Net and Retinex-Net outperform other methods due to the specific structure-awareness illumination smoothness loss term, which makes it more consistent with the image structure. Besides, both as image decomposition networks, the biggest difference between SID-Net and Retinex-Net is that the proposed SID-Net is an unsupervised learning network without any prior training process, while the other one is a supervised learning network based on large-scale training dataset. As for the practical image decomposition performance, the two networks are evenly matched. Even in the reflectance maps, the results of SID-Net are clearer with more details than that of Retinex-Net. In general, as shown in Fig. 4, SID-Net can extract underlying consistent reflectance from the given input image directly, and portray the lightness and shadow of the image in the illumination map. Although without prior training, the SID-Net can achieve the same or even better performance at the image decomposition stage.

Refer to caption
Figure 4: The decomposition results of LIME, JED, Retinex-Net, and SID-Net for Bookshelf in LOL dataset and Balloon in MEF dataset. For each of the example, the first row are the decomposed reflectance, while the second are the illumination maps

5.3 Human Perception User Study

User study was conducted to perform comparisons between state-of-the-art methods in the dimensions of the overall image quality and some unpleasing artifacts through human perceptions. To evaluate the overall image quality, the subjects gave their personal feelings on the scale of 0 to 10, where 0 represents “poor”, 5 means “good”, and 10 is “excellent”. As for the artifacts, there are five artifacts to be selected from (multiple-choice). Those are black edges, blur, overexposure, gray shadow and halo as shown in Fig. 5. Once the subjects observe phenomenon in the images consisting with the description of the artifacts, the corresponding option would be checked. For all the enhanced images obtained by different methods, the subjects move the slider to give the image quality score and select specific artifacts according to their observations. The above procedure was completed on the Amazon Mechanical Turk platform by thirty workers.

Refer to caption

(a) Black Edges

Refer to caption

(b) Blur

Refer to caption

(c) Overexposure

Refer to caption

(d) Gray Shadow

Refer to caption

(e) Halo

Figure 5: Five artifacts examples from (a) Retinex-Net [22] on hill in DICM dataset; (b) JED [17] on tree in DICM dataset; (c) LIME [6] on stormtrooper in LIME dataset; (d) and (e) ExCNet [24] on roof in MEF dataset.

The results of the user study are summarized in graphs shown in Fig. 6 and 7. In Fig. 6, a boxplot is used to exhibit the distributions of the image quality scores for different methods, in which the red solid line represents the median while the blue dotted line represents the mean score. Through observing the median, mean, and the first quartile, it is obvious that the participants showed a strong bias in preference towards our results (median: 6.0, mean: 5.51, quartile: 8.0) when compared to Retinex-Net [22] (median: 3.0, mean: 3.05, quartile: 4.0), gave higher scores when compared to SRIE [4] (median: 5.0, mean: 5.04, quartile: 7.0), JED [17] (median: 5.0, mean: 5.34, quartile: 7.0), PBS [25] (median: 5.0, mean: 5.31, quartile: 7.0) and ExCNet [24] (median: 5.0, mean: 5.41, quartile: 7.0), and had no preference when compared with LIME [6] (median: 6.0, mean: 5.67, quartile: 8.0). The results in Fig. 6 indicate that SID-NISM outperforms most of the state-of-the-art methods in terms of the overall image quality.

Refer to caption
Figure 6: The image quality scores for different methods.

In Fig. 7, a stacked bar graph is adopted to illustrate the average counts of artifacts in the results of different methods, in which each color bar corresponds to the five artifacts in Fig. 5 respectively. Specifically, the average counts are computed by M/(NP)M/(N*P), where MM is the total counts of artifacts in the results of each method, NN is the number of images, PP is the number of subjects. It can be seen that SID-NISM tends to introduce less artifacts (average counts: 0.6) than other methods (average counts: above 0.8) in total. Generally, SID-NISM doesn’t lead too many unexpected artifacts. SRIE [4] would introduce gray shadow into their results, while the results of LIME [6] may be over-enhanced to become overexposure. JED [17] has more blurry cases, while Retinex-Net [22] contains the most black edges in their results. These phenomenons can be also observed in the visual examples of the next subsection.

Refer to caption
Figure 7: The average counts of artifacts in different methods.

5.4 Visual Quality Comparisons

In order to facilitate readers to visually compare the results of different low-light image enhancement approaches, several examples are exhibited in Fig. 8 and  9.

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

(a) Input

Refer to caption

(b) SRIE

Refer to caption

(c) LIME

Refer to caption

(d) JED

Refer to caption

(e) SID-NISM

Figure 8: Compared with the results obtained by SRIE [4], LIME [6] and JED [17] on two natural images: Totoro from LOL dataset and candle from MEF dataset. Below are the enlarged detailed images in the red box. (Best viewed on screen)

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

Refer to caption

(a) Input

Refer to caption

(b) PBS

Refer to caption

(c) Retinex-Net

Refer to caption

(d) ExCNet

Refer to caption

(e) SID-NISM

Figure 9: Compared with the results obtained by PBS [25], Retinex-Net [22] and ExCNet [24] on two natural images: street from LIME dataset and toy from LOL dataset. Below are the enlarged detailed images in the red box. (Best viewed on screen)

In general, it is clear that SID-NISM could brighten up the dark regions of low-light images successfully. More importantly, it can also restore the strong contrast of scene content sufficiently. For example, taking candle in Fig. 8 as input, the contrast between the candle in the foreground and the glass cup in the background is clearer and stronger in the result of SID-NISM compared with other methods. Another example is the toy in Fig. 9. It is obvious that the contrast between the toy’s face and the background is stronger in our result. This property renders the results of our method seem more active and attractive, which is consistent with the user study about the image overall perceptions.

In addition, as discussed in Sect. 5.3, SID-NISM could restore the overall quality of low-light images on the premise of introducing less unexpected artifacts, including preserving clear boundary (glass cup in candle), avoiding gray shadow (the inside of the door in street) and halo (boundary of the door in street). Especially, our results contain less noise such as the sky in street and the white background in toy. Combining with the above intuitive observation and data statistics in the user study, it can be seen that our method is more robust with less artifacts, which makes our framework more competitive and practical in both people’s daily use and commercial cases.

5.5 Objective Evaluation Indexes

GE CE GMI GMG NIQE PSNR SSIM
SRIE 6.5584 18.4951 56.0559 6.2808 3.8777 12.8554 0.5298
LIME 7.2652 19.8994 93.2529 11.4664 4.2036 17.1717 0.6355
JED 6.5296 19.0923 73.9586 4.6584 3.3546 14.9944 0.6325
PBS 7.0828 20.2996 84.7626 10.8144 4.4976 16.3765 0.6375
Retinex-Net 6.9750 21.0380 110.2008 12.8826 5.7127 14.4061 0.4932
ExCNet 7.0548 18.7087 87.4292 10.2022 4.4485 14.9694 0.601
SID-NISM w/o R\mathcal{L}_{R} and N\mathcal{L}_{N} 7.1978 18.8489 99.9323 11.1822 4.4493 17.0313 0.5926
SID-NISM 7.2499 20.7416 103.3472 9.9731 3.8579 17.7576 0.6633
Reference 7.1949 21.0804 110.8416 7.1368 3.5593 / 1
Table 1: Quantitative comparison on LOL and MEF datasets in terms of different metrics. GE, CE, GMI, GMG should be close to the reference images. Lower NIQE, higher PSNR and SSIM indicate better image quality. Top2 results are highlighted in bold (Top1-Top2).
bicycle boat bottle bus car cat chair dog motor person mAP
Original 42.36 8.34 21.38 42.31 30.74 20.44 34.77 25.53 33.12 36.66 31.54
SRIE 44.75 10.76 27.26 54.01 30.80 21.28 36.96 26.27 34.02 38.33 33.92
LIME 47.56 9.25 24.26 53.24 32.72 20.40 32.25 28.29 32.50 39.98 33.98
JED 41.50 8.53 21.42 46.53 23.72 20.46 31.36 30.24 31.77 33.83 29.28
PBS 47.33 8.31 23.85 50.43 33.12 17.65 33.44 26.81 33.19 38.29 33.20
Retinex-Net 34.59 5.70 17.80 38.11 20.99 16.18 27.57 6.14 15.53 27.90 23.15
ExCNet 39.24 8.86 19.46 55.37 32.99 20.68 32.00 30.13 32.49 39.04 32.84
SID-NISM 47.64 9.69 18.26 51.36 32.02 17.19 36.48 28.53 32.75 39.58 33.39
Table 2: The performance (mAP: %) of object detection on real low-light images and their enhanced versions by low-light image enhancement methods.

In addition to the above subjective experiments, there are many indexes with or without reference images that can be used to evaluate the quality of the enhanced images objectively. In this section, we take use of gray entropy (GE), color entropy (CE, sum of the entropy of RGB channels), gray mean illumination (GMI), gray mean gradient (GMG), NIQE [15], PSNR, and SSIM [21] to compare different low-light image enhancement methods. Generally, the first four indexes could reflect the gap between the enhanced results and the reference high-light images from four aspects including the information in gray and color channels, the illumination degree and the image gradient magnitude. NIQE assesses the overall naturalness of the enhanced results and a lower value roughly corresponds to a higher overall naturalness. PSNR could verify the performance of denoising algorithms and a higher value usually indicates a better image quality. SSIM measures the structure similarity between the enhanced results and the reference images, which is in the range of [0, 1]. It should be noted that these indexes can only reflect the image quality in specific aspects, which may be not completely consistent with the evaluation results given by the human visual system.

The quantitative results over LOL and MEF datasets (the two datasets with reference normal-light images) are reported in Table 1. Obviously, the proposed SID-NISM could achieve better performance compared with existing state-of-the-art methods in terms of the image information, the overall quality and the noise condition. It is worth mentioning that we also compare the performance of our framework with and without R\mathcal{L}_{R} and N\mathcal{L}_{N} to quantitatively evaluate the merit brought by the two novel reflectance and noise terms in the loss function of SID-Net. It can be seen that they not only help improve the image color and gradient information of the enhanced results greatly, but also decrease the noise influence and promote the overall image quality.

6 Object Detection Test

We also take a step further to look into the potential of low-light image enhancement in improving the performance of one popular computer vision task, object detection. To do so, tests were performed on YOLOv2 platform on real low-light images and their enhanced versions by our proposed framework and other state-of-the-art methods. Here we directly used the pre-trained VOC2007+2012 model without any re-training or fine-tuning as we would like to observe the outcome of the originally optimized model when given low-light and enhanced images. Experiments were conducted on Exclusively Dark Image Dataset (ExDark), which is consist of low-light images taken in 10 different low-light environments with corresponding 12 image classes and object level annotations.

Table 2 exhibits the mean of Average Precision (mAP) over 10 classes when detecting different versions of images. Obviously, compared to detecting objects on the original low-light images directly, the performance of the object detection model would be improved by preprocessing the inputs through our proposed low-light image enhancement method from 31.54% to 33.39% in terms of mAP. Among existing methods, LIME reaches the best mAP (33.98%) while SID-NISM shows its competitive power (within 1%) in improving the performance of downstream tasks. This test illustrates the feasibility of incorporating image enhancement as a support for practical applications.

7 Conclusion

In this paper, a self-supervised low-light image enhancement method called SID-NISM is proposed, which can first decompose given input images in an unsupervised way without relying on any external examples or prior training, and then enhance decomposed illumination map by a well-designed nonlinear illumination mapping function. The proposed scheme is concise yet powerful. Experimental results show that, even though our method does not need any support data or prior knowledge, it can produce visually pleasing enhancement results with less unexpected artifacts as well as a good representation of image decomposition.

References

  • [1] Vladimir Bychkovsky, Sylvain Paris, Eric Chan, and Frédo Durand. Learning photographic global tonal adjustment with a database of input/output image pairs. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 97–104. IEEE, 2011.
  • [2] Kostadin Dabov, Alessandro Foi, Vladimir Katkovnik, and Karen Egiazarian. Image denoising by sparse 3-d transform-domain collaborative filtering. IEEE Transactions on image processing, 16(8):2080–2095, 2007.
  • [3] Kevin Dale, Micah K Johnson, Kalyan Sunkavalli, Wojciech Matusik, and Hanspeter Pfister. Image restoration using online photo collections. In IEEE 12th International Conference on Computer Vision (ICCV), pages 2217–2224. IEEE, 2009.
  • [4] Xueyang Fu, Delu Zeng, Yue Huang, Xiao-Ping Zhang, and Xinghao Ding. A weighted variational model for simultaneous reflectance and illumination estimation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2782–2790. IEEE, 2016.
  • [5] Roger Grosse, Micah K Johnson, Edward H Adelson, and William T Freeman. Ground truth dataset and baseline evaluations for intrinsic image algorithms. In IEEE 12th International Conference on Computer Vision (ICCV), pages 2335–2342. IEEE, 2009.
  • [6] Xiaojie Guo, Yu Li, and Haibin Ling. Lime: Low-light image enhancement via illumination map estimation. IEEE Transactions on Image Processing, 26(2):982–993, 2016.
  • [7] Daniel J Jobson, Zia-ur Rahman, and Glenn A Woodell. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Transactions on Image processing, 6(7):965–976, 1997.
  • [8] Daniel J Jobson, Zia-ur Rahman, and Glenn A Woodell. Properties and performance of a center/surround retinex. IEEE Transactions on Image Processing, 6(3):451–462, 1997.
  • [9] Sing Bing Kang, Ashish Kapoor, and Dani Lischinski. Personalization of image enhancement. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages 1799–1806. IEEE, 2010.
  • [10] Edwin H Land and John J McCann. Lightness and retinex theory. Josa, 61(1):1–11, 1971.
  • [11] Chulwoo Lee, Chul Lee, and Chang-Su Kim. Contrast enhancement based on layered difference representation of 2d histograms. IEEE Transactions on Image Processing, 22(12):5372–5384, 2013.
  • [12] Mading Li, Jiaying Liu, Wenhan Yang, Xiaoyan Sun, and Zongming Guo. Structure-revealing low-light image enhancement via robust retinex model. IEEE Transactions on Image Processing, 27(6):2828–2841, 2018.
  • [13] Kede Ma, Kai Zeng, and Zhou Wang. Perceptual quality assessment for multi-exposure image fusion. IEEE Transactions on Image Processing, 24(11):3345–3356, 2015.
  • [14] Wenye Ma, Stanley Osher, and Selim Esedoglu. A tv bregman iterative model of retinex theory. Inverse Problems and Imaging, 6(4):697–708, 2012.
  • [15] Anish Mittal, Rajiv Soundararajan, and Alan C Bovik. Making a “completely blind” image quality analyzer. IEEE Signal Processing Letters, 20(3):209–212, 2012.
  • [16] Michael K Ng and Wei Wang. A total variation model for retinex. SIAM Journal on Imaging Sciences, 4(1):345–365, 2011.
  • [17] Xutong Ren, Mading Li, Wen-Huang Cheng, and Jiaying Liu. Joint enhancement and denoising method via sequential decomposition. In IEEE International Symposium on Circuits and Systems (ISCAS), pages 1–5. IEEE, 2018.
  • [18] Liang Shen, Zihan Yue, Fan Feng, Quan Chen, Shihao Liu, and Jie Ma. Msr-net: Low-light image enhancement using deep convolutional network. arXiv preprint arXiv:1711.02488, 2017.
  • [19] Ruixing Wang, Qing Zhang, Chi-Wing Fu, Xiaoyong Shen, Wei-Shi Zheng, and Jiaya Jia. Underexposed photo enhancement using deep illumination estimation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 6849–6857. IEEE, 2019.
  • [20] Shuhang Wang, Jin Zheng, Hai-Miao Hu, and Bo Li. Naturalness preserved enhancement algorithm for non-uniform illumination images. IEEE Transactions on Image Processing, 22(9):3538–3548, 2013.
  • [21] Zhou Wang, Alan C Bovik, Hamid R Sheikh, and Eero P Simoncelli. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13(4):600–612, 2004.
  • [22] Chen Wei, Wenjing Wang, Wenhan Yang, and Jiaying Liu. Deep retinex decomposition for low-light enhancement. arXiv preprint arXiv:1808.04560, 2018.
  • [23] Lu Yuan and Jian Sun. Automatic exposure correction of consumer photographs. In European Conference on Computer Vision (ECCV), pages 771–785. Springer, 2012.
  • [24] Lin Zhang, Lijun Zhang, Xiao Liu, Ying Shen, Shaoming Zhang, and Shengjie Zhao. Zero-shot restoration of back-lit images using deep internal learning. In ACM Multimedia Conference on Multimedia Conference (ACMMM), pages 1623–1631. ACM, 2019.
  • [25] Qing Zhang, Ganzhao Yuan, Chunxia Xiao, Lei Zhu, and Wei-Shi Zheng. High-quality exposure correction of underexposed photos. In ACM Multimedia Conference on Multimedia Conference (ACMMM), pages 582–590. ACM, 2018.
  • [26] Yonghua Zhang, Jiawan Zhang, and Xiaojie Guo. Kindling the darkness: A practical low-light image enhancer. In ACM Multimedia Conference on Multimedia Conference (ACMMM), pages 1632–1640. ACM, 2019.