Low-Light Image Fusion Techniques
Low-Light Image Fusion Techniques
8, AUGUST 2015 1
…
Generator
Evaluator Combiner
images into the final enhanced result. Based on our frame- because a global processing can not ensure all local areas be
work, we propose a dual-exposure fusion method. Specifically, well enhanced.
we first employ the illumination estimation techniques to
build the weight matrix for image fusion. Then we derives B. Local Enhancement Algorithms
our camera response model based on observation. Next, we
find the optimal exposure for our camera response model By making use of spatial information directly, local en-
to generate the synthetic image that is well-exposed in the hancement can achieve better results and become the main-
regions where the original image is under-exposed. Finally, we stream of recent techniques. Local histogram equalization
obtain the enhanced results by fusing the input image with the [36], [37] adopts the sliding window strategy to perform
synthetic image using the weight matrix. Experiments on five HE locally. Based on the observation that the inverted low-
challenging datasets are conducted to reveal the advantages of light images are closed to hazy images, dehazing techniques
our method in comparison with other state-of-the-art methods. are borrowed to solve low-light image enhancement in some
methods [38], [39]. However, the basic models of above
methods are lacking in physical explanation [14]. To provide
II. R ELATED W ORK a physical meaningful model for image enhancement, Retinex
In general, image enhancement techniques can improve the theory assumes that the amount of light reaching observers
subjective visual quality of input images and support the can be decomposed into two parts: illumination and scene
extraction of valuable information for some computer vision reflection. Most Retinex-based methods get enhanced results
techniques [29], [16]. Low-light image enhancement, as one by removing the illumination part [18], [20], [15] while others
of enhancement techniques, can reveal the information of the [40], [16], [14] keep a portion of the illumination to preserve
under-exposed regions in an image. Broadly speaking, existing naturalness. Fu et al. [41] adjust the illumination components
low-light image enhancement techniques can be divided into by fusing it with two enhanced illumination. As far as we
two major categories: global enhancement and local enhance- know, there is no multi-exposure fusion method for this task
ment. since lowlight enhancement problem usually takes a single
image as input.
B. Multi-Exposure Generator
As aforementioned, images taken with different exposures (e) µ = 1.25 (f) µ = 1.75 (g) µ = 2.25
are correlated. Multi-Exposure Generator maps the input im-
Fig. 3. The enhanced results using different µ.
age into multi-exposure images according to the specified ex-
posure ratio set. The key part of the Multi-Exposure Generator
is the camera response model used to find an appropriate BTF A. Dual-Exposure Evaluator
for mapping. Given an exposure ratio ki and a BTF g, we can
The design of W is key to obtaining an enhancement
map the input image P to the i-th image in the exposure set
algorithm that can enhance the low contrast of under-exposed
as
regions while the contrast in well-exposed regions preserved.
Pi = g(P, ki ). (1)
We need to assign big weight values to well-exposed pixels
C. Multi-Exposure Evaluator and small weight values to under-exposed pixels. Intuitively,
the weight matrix is positively correlated with the scene illu-
To estimate the wellness of each pixel in the generated
mination. Since highly illuminated regions have big possibility
images. The Multi-Exposure Evaluator takes in an image and
of being well-exposed, they should be assign with big weight
outputs a weight matrix that indicates the wellness of each
values to preserve their contrast. We calculate the weight
pixel. The weight matrix is nonuniform for all pixels: the well-
matrix as
exposed pixels are given a big weight while the poor-exposed
Ŵ = Tµ (5)
pixels are given a small weight. After all images are evaluated,
the output matrices are pixel-wise normalized to ensure their where T is the scene illumination map and µ is a parameter
summation equals one for each pixel as controlling the enhance degree. When µ = 0, the resulting
N
X R is equal to P, i.e. , no enhancement is performed. When
f i = Wi
W Wi , (2) µ = 1, both the under-exposed pixels and well-exposed pixels
i=1 are enhanced. When µ > 1, pixels may get saturated and the
where is the element-wise division, and Wi and W f i are resulting R suffers from detail loss. As shown in Fig. 3. In
the i-th matrix before and after normalization, respectively. order to perform enhancement while preserve the well-exposed
regions, we set µ to 0.5. The scene illumination map T is
D. Multi-Exposure Combiner estimated by solving an optimization problem.
To obtain an image with all pixel well-exposed, we can 1) Optimization Problem: The lightness component can be
simply fuse these images based on the weight matrix as used as an estimation of scene illumination. We adopt the
N
lightness component as the initial estimation of illumination:
X
Rc = f i ◦ Pc ,
W i (3) L(x) = max Pc (x) (6)
i=1 c∈{R,G,B}
where c is the index of three color channels and R is for each individual pixel x. Ideal illumination should has local
the enhanced result. Other fusion techniques like multi-scale consistency for the regions with similar structures. In other
fusion [42] and Boosting Laplacian Pyramid fusion [43] can words, T should keep the meaningful structures of the image
also be used to obtain a better fusion results. and remove the textural edges. As in [14], we refine T by
IV. D UAL -E XPOSURE F USION A LGORITHM solving the following optimization equation:
In this section, we use the proposed framework to design a min ||T − L||22 + λ||M ◦ ∇T||1 , (7)
T
low-light image enhancement algorithm. To reduce complex-
ity, we only generate one image with appropriate exposure and where || ∗ ||2 and || ∗ ||1 are the `2 and `1 norm, respectively.
obtain the enhanced result by fusing the input image and the The first order derivative filter ∇ contains ∇h T (horizontal)
generated one. Based on our framework, the fused image is and ∇v T (vertical). M is the weight matrix and λ is the
defined as coefficient. The first term of this equation is to minimize the
difference between the initial map L and the refined map T,
Rc = Ŵ ◦ Pc + (1 − Ŵ) ◦ g(Pc , k̂). (4) while the second term maintains the smoothness of T.
The enhancement problem can be divided into three parts: The design of M is important for the illumination map
the determination of multi-exposure Evaluator (Ŵ), multi- refinement. A major edge in a local window contributes
exposure generator (g), and multi-exposure sampler (k̂). In more similar-direction gradients than textures with complex
the following subsections, we solve them one by one. patterns [44]. Therefore, The weight in a window that contains
JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 4
B. Dual-Exposure Generator
In this section, we present a camera response model to
implement the Multi-Exposure Generator. A camera response
model consists of two parts: Camera Response Function (CRF)
Fig. 5. Observation. From left to right: An under-exposure image P0 , apply
model and BTF model. The parameters of CRF model is multiplication αP0 , apply gamma function (αP0 )γ and the well-exposure
determined only by camera while that of BTF model is deter- image P1 under the same scene. The histograms of red, green and blue color
mined by camera and exposure ratio. In this subsection, we channels are plotted above the corresponding image respectively.
JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 5
Fig. 6. Examples of multi-exposure image sets and their entropy values. The red bar on the left of each image shows the exponential of its entropy value.
Two CRF models can be derived from two cases of Eq. 13. arithmetic mean) since it has the same BTF model parameters
When γ = 1, the CRF model is a power function and the (β and γ) with all three color channels, as shown in Eq. 19.
BTF model is a simple linear function. As some camera q
manufacturers design f to be a gamma curve, it can fit B0 := 3 Q0r ◦ Q0g ◦ Q0b
these cameras perfectly. When γ 6= 1, the CRF model is a q
= 3 (βQγr ) ◦ (βQγg ) ◦ (βQγb ) = β( 3 Qr ◦ Qg ◦ Qb )γ
p
two-parameter function and the BTF model is a non-linear
function. Since the BTF is non-linear for most cameras, we = βBγ .
mainly consider the case of γ 6= 1. Our BTF g is solved as (19)
The visibility of a well-exposed image is higher than that
a a
g(P, k) = eb(1−k ) P(k ) . (16) of an under/over-exposed image and it can provide a richer
information for human. Thus, the optimal k should provide
where β and γ are two model parameters that can be calculated the largest amount of information. To measure the amount of
from camera parameters a, b and exposure ratio k. We assume information, we employ the image entropy which is defined
that no information about the camera is provided and use a as
N
fixed camera parameters (a = −0.3293, b = 1.1258) that can
X
H(B) = − pi · log2 pi , (20)
fit most cameras. i=1
First, we exclude the well-exposed pixels and obtain an Since the image entropy increases first and then decreases with
image that is globally under-exposed. We simply extract the the increase of the exposure ratio, k̂ can be solved by one-
low illuminated pixels as dimensional minimizer. To improve the calculation efficiency,
we resize the input image to 50 × 50 when optimizing k.
Q = {P(x)|T(x) < 0.5}, (17)
V. E XPERIMENTS
where Q contains only the under-exposed pixels.
To evaluate the performance of our method, we com-
The brightness of the images under different exposures pare it with several state-of-the-art methods, including Multi
changes significantly while the color is basically the same. Scale Retinex with Color Restoration (MSRCR) [45], Natural-
Therefore, we only consider the brightness component when ness Preserved Enhancement algorithm (NPE) [16], dehazing
estimating k. The brightness component B is defined as the based method (Dong) [38], Multi-deviation Fusion method
geometric mean of three channel: (MF) [41], Illumination Estimation based method (LIME)
p [14] and Simultaneous Reflection and Illumination Estimation
B := 3
Qr ◦ Qg ◦ Qb , (18) (SRIE) [46]. We test those methods on hundreds of low-light
images from five public datasets: VV [47], LIME-data [14],
where Qr , Qg and Qb are the red, green and blue channel of NPE-data and its extension (NPE, NPE-ex1, NPE-ex2 and
the input image Q respectively. We use the geometric mean NPE-ex3) [16], DICM [3], and MEF [48]. The datasets are
instead of other definitions (e.g. arithmetic mean and weighted briefly introduced as follows:
JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 6
Fig. 7. Comparison of lightness distortion. The odd rows show the original image and the results of various enhancement methods, and the even rows show
the visualization of each method’s lightness distortion (RD).
TABLE I
Q UANTITATIVE MEASUREMENT RESULTS OF LIGHTNESS DISTORTION (LOE).
VV1 . This dataset is collected by Vassilios Vonikakis in his In order to maintain the fairness of the comparison, all the
daily life to provide the most challenging cases for enhance- codes are in Matlab and all the experiments are conducted on
ment. Each image in the dataset has a part that is correctly a PC running Windows 10 OS with 64G RAM and 3.4GHz
exposed and a part that is severely under/over-exposed. A CPU (GPU acceleration is not used). The parameters of our
good enhancement algorithm should enhance the under/over- enhancement algorithm are fixed in all experiments: λ = 1,
exposed regions while not affect the correctly exposed one. = 0.001, µ = 1/2, and the size of local window ω(x) is 5.
LIME-data2 . This dataset contains 10 low-light images The most time-consuming part of our algorithm is illumination
used in [14]. map optimization. We employ the multi-resolution precondi-
NPE3 . This dataset contains 85 low-light images down- tioned conjugate gradient solver (O(N )) to solve it efficiently.
loaded from Internet. NPE-data contains 8 outdoor nature
scene images which are used in [16]. NPE-ex1, NPE-ex2 and A. Lightness Distortion
NPE-ex3 are three supplementary datasets including cloudy We use lightness order error (LOE) to objectively measure
daytime, daybreak, nightfall and nighttime scenes. the lightness distortion of enhanced results. LOE is defined as
DICM4 . It contains 69 captured images from commercial m
digital cameras collected by [3]. 1 X
LOE = RD(x) (22)
MEF5 . This dataset was provided by [48]. It contains m x=1
17 high-quality image sequences including natural sceneries, where RD(x) is the relative order difference of the lightness
indoor and outdoor views and man-made architectures. Each between the original image P and its enhanced version P 0 for
image sequence has several multi-exposure images, we select pixel x, which is defined as follows:
one of poor-exposed images as input to perform evaluation. m
X
RD(x) = U L(x), L(y) ⊕ U L0 (x), L0 (y) , (23)
1 https://sites.google.com/site/vonikakis/datasets y=1
2 http://cs.tju.edu.cn/orgs/vision/∼xguo/LIME.htm
3 http://blog.sina.com.cn/s/blog a0a06f190101cvon.html where m is the pixel number, ⊕ stands for the exclusive-
4 http://mcl.korea.ac.kr/projects/LDR/LDR TEST IMAGES DICM.zip or operator, L(x) and L0 (x) are the lightness component at
5 https://ece.uwaterloo.ca/∼k29ma/ location x of the input images and the enhanced images,
JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 7
Fig. 8. Comparison of contrast distortion. The loss of visible contrast is marked in green, the amplification of invisible contrast is marked in blue, and the
reversal of visible contrast is marked in red. Different shades of color represent different degrees of distortion.
TABLE II
Q UANTITATIVE MEASUREMENT RESULTS OF VISUAL INFORMATION FIDELITY (VIF).
respectively. The function U (p, q) returns 1 if p >= q, 0 proposed method obtains the most realistic results with the
otherwise. least distortion.
As suggested in [14], [16], down-sampling is used to reduce
the complexity of computing LOE. We notice that LOE C. Visual Information Distortion
may change significantly when an image is down-sampled to
different sizes since RD will increase as the pixel number m To measure the distortion of visual information, we employ
increases. Therefore, we down-sample all images to a fixed Visual Information Fidelity (VIF) [50] in reverse mode. As a
size. Specifically, we collect 100 rows and columns evenly to full reference image quality assessment index, VIF models the
form a 100 × 100 down-sampled image. quality assessment problem as a information fidelity criterion
As shown in Table I, our algorithm outperforms the others in that quantifies the mutual information between the reference
all datasets. This means that our algorithm can maintain the image C and the distorted image F relative to the information
naturalness of images well. We also provide a visualization of C extracted by the HVS. VIF measure is given by
of lightness distortion on two cases in Fig. 7, from which, I(C; F )
we can find our results have the smallest lightness distortion. V IF = , (24)
I(C; E)
The results of MSRCR lose the global lightness order and
suffer from severe lightness distortion. Although the results of where E is the image that the HVS perceives. The mutual
LIME is visually pleasant, they are full of lightness distortion. information I(C; F ) and I(C; E) represent the information
The results of Dong, NPE, MF and SRIE can only retain the that could be extracted by the brain in the reference and the
lightness order in the well-exposed regions. test images respectively.
Like most of full reference image quality assessment meth-
ods, VIF were designed for and tested on degraded images.
B. Contrast Distortion The normal version of VIF treats the original image as the
As aforementioned, the image that only differ in exposures reference image and the image outputted by algorithm as the
can be used as a reference for evaluating the accuracy of degraded image. For image enhancement problem, however,
enhanced results. DRIM (Dynamic Range Independent Metric) the original image is the degraded one. Therefore, we employ
[49] can measure the distortion of image contrast without VIF in reverse mode by specifying the enhanced version of the
the interference of change in image brightness. We use it image as the reference and the original image as the degraded
to visualize the contrast difference between the enhanced image. VIF provides consistently high value of correlation be-
result and the reference image. As shown in Fig. 8, the tween subjective MOS (Mean Opinion Score) and its scores, as
JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 8
shown in [51]. Besides, it is suitable for automating evaluation images), it may make the results look whitish. Besides, in
process of nonlinear image enhancement algorithms [52]. some dark show halo artifacts around sharp edges. The results
As shown in Table II, our algorithm outperforms the others of MSRCR shows severe halo artifacts around sharp edges and
in all datasets. This means that our algorithm can maintain the obvious noise in very dark areas (see the two astronauts) and
visual information of images well. suffer from detail loss in some bright areas (see the Christmas
hat). The results of Dong is noisy and full of bold edges
D. Time Cost making it look like exaggerated art paintings. The results
of LIME is so bright that many bright areas are saturated.
Fig. 10 gives the comparison among different methods Also, the noise in dark areas are amplified and de-noising
in terms of time cost. Although SRIE and NPE produce method are therefore required to obtain better results. MF may
small distortion, they are quite time-consuming. Our method introduce color over-enhancement (see the ground beneath the
achieves the smallest distortion than others with an acceptable astronauts’ feet) and SRIE may produce slight halo effects in
time cost. some edges (see the Christmas hat).
2.5
LIME This highlights a direction for future work. To avoid the
MF
Ours over-enhancement due to the ignorance of the scene content,
Run time (ms)
2
semantic understanding is required. With further refinement,
1.5
we might employ the deep learning techniques to estimate the
1 illumination map.
0.5
Besides, we only use two images to obtain the enhanced
0
result. The over-exposure problem is remain unsolved. Im-
0 0.2 0.4 0.6 0.8 1 1.2 1.4
Image Size (width × height)
1.6 1.8 2
106
ages with smaller exposures than the input image should be
considered in our framework to obtain a better result. We will
Fig. 10. Time comparison among different methods with varying image sizes address this problem as future work.
JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 9
[21] K. Panetta, S. Agaian, Y. Zhou, and E. J. Wharton, “Parameterized [46] X. Fu, D. Zeng, Y. Huang, X.-P. Zhang, and X. Ding, “A weighted vari-
logarithmic framework for image enhancement,” IEEE Transactions on ational model for simultaneous reflectance and illumination estimation,”
Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 41, no. 2, pp. in Proceedings of the IEEE Conference on Computer Vision and Pattern
460–473, 2011. Recognition, 2016, pp. 2782–2790.
[22] K. A. Panetta, E. J. Wharton, and S. S. Agaian, “Human visual system- [47] V. Vonikakis, R. Kouskouridas, and A. Gasteratos, “On the evaluation
based image enhancement and logarithmic contrast measure,” IEEE of illumination compensation algorithms,” Multimedia Tools and Appli-
Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), cations, pp. 1–21, 2017.
vol. 38, no. 1, pp. 174–188, 2008. [48] K. Ma, K. Zeng, and Z. Wang, “Perceptual quality assessment for
[23] Y. Wang, S. Zhuo, D. Tao, J. Bu, and N. Li, “Automatic local exposure multi-exposure image fusion,” IEEE Transactions on Image Processing,
correction using bright channel prior for under-exposed images,” Signal vol. 24, no. 11, pp. 3345–3356, 2015.
Processing, vol. 93, no. 11, pp. 3227–3238, 2013. [49] T. O. Aydin, R. Mantiuk, K. Myszkowski, and H.-P. Seidel, “Dynamic
[24] L. Yuan and J. Sun, “Automatic exposure correction of consumer range independent image quality assessment,” ACM Transactions on
photographs,” Computer Vision–ECCV 2012, pp. 771–785, 2012. Graphics (TOG), vol. 27, no. 3, p. 69, 2008.
[25] G. Deng, “A generalized unsharp masking algorithm,” IEEE transactions [50] H. R. Sheikh and A. C. Bovik, “Image information and visual quality,”
on Image Processing, vol. 20, no. 5, pp. 1249–1261, 2011. IEEE Transactions on image processing, vol. 15, no. 2, pp. 430–444,
[26] Q. Shan, J. Jia, and M. S. Brown, “Globally optimized linear windowed 2006.
tone mapping,” IEEE transactions on visualization and computer graph- [51] K. Gu, G. Zhai, W. Lin, and M. Liu, “The analysis of image contrast:
ics, vol. 16, no. 4, pp. 663–675, 2010. From quality assessment to automatic enhancement,” IEEE transactions
[27] J.-H. Hong, S.-B. Cho, and U.-K. Cho, “A novel evolutionary approach on cybernetics, vol. 46, no. 1, pp. 284–297, 2016.
to image enhancement filter design: method and applications,” IEEE [52] P. A. Kumar and P. Sankaran, “Visual information fidelity in evaluating
Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), retinex enhancement algorithms,” in Communications and Signal Pro-
vol. 39, no. 6, pp. 1446–1457, 2009. cessing (ICCSP), 2014 International Conference on. IEEE, 2014, pp.
[28] S. T. McHugh, Understanding Photography. Cambridge in Colour, 167–171.
2016.
[29] H. Ibrahim and N. S. P. Kong, “Brightness preserving dynamic histogram
equalization for image contrast enhancement,” IEEE Transactions on
Consumer Electronics, vol. 53, no. 4, pp. 1752–1758, 2007.
[30] A. Beghdadi and A. Le Negrate, “Contrast enhancement technique based
on local detection of edges,” Computer Vision, Graphics, and Image
Processing, vol. 46, no. 2, pp. 162–174, 1989.
[31] E. Peli, “Contrast in complex images,” JOSA A, vol. 7, no. 10, pp.
2032–2040, 1990.
[32] R. Gonzalez and R. Woods, “Digital image processing: Pearson prentice
hall,” Upper Saddle River, NJ, 2008.
[33] A. K. Jain, Fundamentals of digital image processing. Prentice-Hall,
Inc., 1989.
[34] C. Wang and Z. Ye, “Brightness preserving histogram equalization with
maximum entropy: a variational perspective,” IEEE Transactions on
Consumer Electronics, vol. 51, no. 4, pp. 1326–1334, 2005.
[35] A. M. Reza, “Realization of the contrast limited adaptive histogram
equalization (clahe) for real-time image enhancement,” Journal of VLSI
signal processing systems for signal, image and video technology,
vol. 38, no. 1, pp. 35–44, 2004.
[36] J. A. Stark, “Adaptive image contrast enhancement using generalizations
of histogram equalization,” IEEE Transactions on image processing,
vol. 9, no. 5, pp. 889–896, 2000.
[37] M. Abdullah-Al-Wadud, M. H. Kabir, M. A. A. Dewan, and O. Chae,
“A dynamic histogram equalization for image contrast enhancement,”
IEEE Transactions on Consumer Electronics, vol. 53, no. 2, 2007.
[38] X. Dong, G. Wang, Y. Pang, W. Li, J. Wen, W. Meng, and Y. Lu, “Fast
efficient algorithm for enhancement of low lighting video,” in 2011 IEEE
International Conference on Multimedia and Expo. IEEE, 2011, pp.
1–6.
[39] L. Li, R. Wang, W. Wang, and W. Gao, “A low-light image enhancement
method for both denoising and contrast enlarging,” in Image Processing
(ICIP), 2015 IEEE International Conference on. IEEE, 2015, pp. 3730–
3734.
[40] X. Fu, D. Zeng, Y. Huang, X.-P. Zhang, and X. Ding, “A weighted vari-
ational model for simultaneous reflectance and illumination estimation,”
in Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition, 2016, pp. 2782–2790.
[41] X. Fu, D. Zeng, Y. Huang, Y. Liao, X. Ding, and J. Paisley, “A
fusion-based enhancing method for weakly illuminated images,” Signal
Processing, vol. 129, pp. 82–96, 2016.
[42] T. Mertens, J. Kautz, and F. Van Reeth, “Exposure fusion: A simple and
practical alternative to high dynamic range photography,” in Computer
Graphics Forum, vol. 28, no. 1. Wiley Online Library, 2009, pp. 161–
171.
[43] J. Shen, Y. Zhao, S. Yan, X. Li et al., “Exposure fusion using boosting
laplacian pyramid.” IEEE transactions on cybernetics, vol. 44, no. 9,
pp. 1579–1590, 2014.
[44] L. Xu, Q. Yan, Y. Xia, and J. Jia, “Structure extraction from texture via
relative total variation,” ACM Transactions on Graphics (TOG), vol. 31,
no. 6, p. 139, 2012.
[45] A. B. Petro, C. Sbert, and J.-M. Morel, “Multiscale retinex,” Image
Processing On Line, pp. 71–88, 2014.