zero-reference deep curve estimation for low-light image ... · dce, namely le-curve, dce-net, and...

9
Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement Chunle Guo * Tianjin University [email protected] Chongyi Li * City University of Hong Kong [email protected] Jichang Guo Tianjin University [email protected] Chen Change Loy Nanyang Technological University [email protected] Junhui Hou City University of Hong Kong [email protected] Sam Kwong City University of Hong Kong [email protected] Runmin Cong Beijing Jiaotong University [email protected] Abstract The paper presents a novel method, Zero-Reference Deep Curve Estimation (Zero-DCE), which formulates light enhancement as a task of image-specific curve estimation with a deep network. Our method trains a lightweight deep network, DCE-Net, to estimate pixel-wise and high-order curves for dynamic range adjustment of a given image.The curve estimation is specially designed, considering pixel value range, monotonicity, and differentiability. Zero-DCE is appealing in its relaxed assumption on reference images, i.e., it does not require any paired or unpaired data dur- ing training. This is achieved through a set of carefully formulated non-reference loss functions, which implicitly measure the enhancement quality and drive the learning of the network. Our method is efficient as image enhance- ment can be achieved by an intuitive and simple nonlinear curve mapping. Despite its simplicity, we show that it gen- eralizes well to diverse lighting conditions. Extensive ex- periments on various benchmarks demonstrate the advan- tages of our method over state-of-the-art methods qualita- tively and quantitatively. Furthermore, the potential ben- efits of our Zero-DCE to face detection in the dark are discussed. Code and model will be available at https: //github.com/Li-Chongyi/Zero-DCE. 1. Introduction Many photos are often captured under suboptimal light- ing conditions due to inevitable environmental and/or tech- nical constraints. These include inadequate and unbalanced * The authors contributed equally to this work. Corresponding author. (a) Raw (b) Zero-DCE (c) Wang et al. [24] (d) EnlightenGAN [10] Figure 1. Visual comparisons on a typical low-light image. The proposed Zero-DCE achieves visually pleasing result in terms of brightness, color, contrast, and naturalness, while existing meth- ods either fail to cope with the extreme back light or generate color artifacts. In contrast to other deep learning-based methods, our ap- proach is trained without any reference image. lighting conditions in the environment, incorrect placement of objects against extreme back light, and under-exposure during image capturing. Such low-light photos suffer from compromised aesthetic quality and unsatisfactory transmis- sion of information. The former affects viewers’ experience while the latter leads to wrong message being communi- cated, such as inaccurate object/face recognition. In this study, we present a novel deep learning-based method, Zero-Reference Deep Curve Estimation (Zero- 4321 arXiv:2001.06826v1 [cs.CV] 19 Jan 2020

Upload: others

Post on 02-Aug-2020

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement

Chunle Guo∗

Tianjin [email protected]

Chongyi Li∗

City University of Hong [email protected]

Jichang Guo†

Tianjin [email protected]

Chen Change LoyNanyang Technological University

[email protected]

Junhui HouCity University of Hong Kong

[email protected]

Sam KwongCity University of Hong Kong

[email protected]

Runmin CongBeijing Jiaotong University

[email protected]

Abstract

The paper presents a novel method, Zero-ReferenceDeep Curve Estimation (Zero-DCE), which formulates lightenhancement as a task of image-specific curve estimationwith a deep network. Our method trains a lightweight deepnetwork, DCE-Net, to estimate pixel-wise and high-ordercurves for dynamic range adjustment of a given image.Thecurve estimation is specially designed, considering pixelvalue range, monotonicity, and differentiability. Zero-DCEis appealing in its relaxed assumption on reference images,i.e., it does not require any paired or unpaired data dur-ing training. This is achieved through a set of carefullyformulated non-reference loss functions, which implicitlymeasure the enhancement quality and drive the learningof the network. Our method is efficient as image enhance-ment can be achieved by an intuitive and simple nonlinearcurve mapping. Despite its simplicity, we show that it gen-eralizes well to diverse lighting conditions. Extensive ex-periments on various benchmarks demonstrate the advan-tages of our method over state-of-the-art methods qualita-tively and quantitatively. Furthermore, the potential ben-efits of our Zero-DCE to face detection in the dark arediscussed. Code and model will be available at https://github.com/Li-Chongyi/Zero-DCE.

1. IntroductionMany photos are often captured under suboptimal light-

ing conditions due to inevitable environmental and/or tech-nical constraints. These include inadequate and unbalanced∗The authors contributed equally to this work.†Corresponding author.

(a) Raw (b) Zero-DCE

(c) Wang et al. [24] (d) EnlightenGAN [10]

Figure 1. Visual comparisons on a typical low-light image. Theproposed Zero-DCE achieves visually pleasing result in terms ofbrightness, color, contrast, and naturalness, while existing meth-ods either fail to cope with the extreme back light or generate colorartifacts. In contrast to other deep learning-based methods, our ap-proach is trained without any reference image.

lighting conditions in the environment, incorrect placementof objects against extreme back light, and under-exposureduring image capturing. Such low-light photos suffer fromcompromised aesthetic quality and unsatisfactory transmis-sion of information. The former affects viewers’ experiencewhile the latter leads to wrong message being communi-cated, such as inaccurate object/face recognition.

In this study, we present a novel deep learning-basedmethod, Zero-Reference Deep Curve Estimation (Zero-

4321

arX

iv:2

001.

0682

6v1

[cs

.CV

] 1

9 Ja

n 20

20

Page 2: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

DCE), for low-light image enhancement. It can cope withdiverse lighting conditions including nonuniform and poorlighting cases. Instead of performing image-to-image map-ping, we reformulate the task as an image-specific curve es-timation problem. In particular, the proposed method takesa low-light image as input and produces high-order curvesas its output. These curves are then used for pixel-wise ad-justment on the dynamic range of the input to obtain an en-hanced image. The curve estimation is carefully formulatedso that it maintains the range of the enhanced image andpreserves the contrast of neighboring pixels. Importantly,it is differentiable, and thus we can learn the adjustable pa-rameters of the curves through a deep convolutional neuralnetwork. The proposed network is lightweight and it can beiteratively applied to approximate higher-order curves formore robust and accurate dynamic range adjustment.

A unique advantage of our deep learning-based methodis zero-reference, i.e., it does not require any paired or evenunpaired data in the training process as in existing CNN-based [24, 27] and GAN-based methods [10, 31]. Thisis made possible through a set of specially designed non-reference loss functions including spatial consistency loss,exposure control loss, color constancy loss, and illumina-tion smoothness loss, all of which take into considerationmulti-factor of light enhancement. We show that even withzero-reference training, Zero-DCE can still perform com-petitively against other methods that require paired or un-paired data for training. An example of enhancing a low-light image comprising nonuniform illumination is shownin Fig. 1. Comparing to state-of-the-art methods, Zero-DCEbrightens up the image while preserving the inherent colorand details. In contrast, both CNN-based method [24] andGAN-based EnlightenGAN [10] yield under-(the face) andover-(the cabinet) enhancement.

Our contributions are summarized as follows.1) We propose the first low-light enhancement network thatis independent of paired and unpaired training data, thusavoiding the risk of overfitting. As a result, our methodgeneralizes well to various lighting conditions.2) We design a simple and lightweight deep network that isable to approximate pixel-wise and higher-order curves byiteratively applying itself. Such image-specific curves caneffectively perform mapping within a wide dynamic range.3) We show the potential of training a deep image enhance-ment model in the absence of reference images throughtask-specific non-reference loss functions that indirectlyevaluate enhancement quality.

The proposed Zero-DCE method supersedes state-of-the-art performance both in qualitative and quantitative met-rics. More importantly, it is capable of improving high-levelvisual tasks, e.g., face detection, without inflicting highcomputational burden. It is capable of processing imagesin real-time (about 500 FPS for images of size 640×480×3

on GPU) and takes only 30 minutes for training.

2. Related WorkConventional Methods. HE-based methods perform lightenhancement through expanding the dynamic range of animage. Histogram distribution of images is adjusted at bothglobal [6, 9] and local levels [23, 13]. There are also variousmethods adopting the Retinex theory [11] that typically de-composes an image into reflectance and illumination. Thereflectance component is commonly assumed to be con-sistent under any lighting conditions; thus, light enhance-ment is formulated as an illumination estimation problem.Building on the Retinex theory, several methods have beenproposed. Wang et al. [25] designed a naturalness- andinformation-preserving method when handling images ofnonuniform illumination; Fu et al. [7] proposed a weightedvariation model to simultaneously estimate the reflectanceand illumination of an input image; Guo et al. [8] first esti-mated a coarse illumination map by searching the maximumintensity of each pixel in RGB channels, then refining thecoarse illumination map by a structure prior; Li et al. [15]proposed a new Retinex model that takes noise into consid-eration. The illumination map was estimated through solv-ing an optimization problem. Contrary to the conventionalmethods that fortuitously change the distribution of imagehistogram or that rely on potentially inaccurate physicalmodels, the proposed Zero-DCE method produces an en-hanced result through image-specific curve mapping. Sucha strategy enables light enhancement on images without cre-ating unrealistic artifacts. Yuan and Sun [29] proposed anautomatic exposure correction method, where the S-shapedcurve for a given image is estimated by a global optimiza-tion algorithm and each segmented region is pushed to itsoptimal zone by curve mapping. Different from [29], ourZero-DCE is a purely data-driven method and considersmultiple light enhancement factors into consideration in thedesign of the non-reference loss functions, and thus enjoysbetter robustness, wider image dynamic range adjustment,and lower computational burden.Data-driven Methods. Data-driven methods are largelycategorized into two branches, namely CNN-based andGAN-based methods. Most CNN-based solutions relyon paired data for supervised training, therefore they areresource-intensive. Often time, the paired data are exhaus-tively collected through automatic light degradation, chang-ing the settings of cameras during data capturing, or syn-thesizing data via image retouching. For example, the LL-Net [16] was trained on data simulated on random Gammacorrection; the LOL dataset [27] of paired low/normal lightimages was collected through altering the exposure timeand ISO during image acquisition; the MIT-Adobe FiveKdataset [3] comprises 5,000 raw images, each of which hasfive corresponding retouched images produced by trained

4322

Page 3: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

experts; Recently, Wang et al. [24] proposed an underex-posed photo enhancement network by estimating the illumi-nation map. This network was trained on paired data whichwas retouched by three experts. Understandably, light en-hancement solutions based on paired data are impractical inmany ways, considering the high cost involved in collectingsufficient paired data as well as the inclusion of factitiousand unrealistic data in training the deep models. Such con-straints are reflected in the poor generalization capabilityof CNN-based methods. Artifacts and color casts are com-monly generated, when these methods are presented withreal-world images of various light intensities.

Unsupervised GAN-based methods have the advantageof eliminating paired data for training. EnlightenGAN [10],an unsupervised GAN-based and pioneer method that learnsto enhance low-light images using unpaired low/normallight data. The network was trained by taking into ac-count elaborately designed discriminators and loss func-tions. However, unsupervised GAN-based solutions usuallyrequire careful selection of unpaired training data.

The proposed Zero-DCE is superior to existing data-driven methods in three aspects. First, it explores a newlearning strategy, i.e., one that requires zero reference,hence eliminating the need for paired and unpaired data.Second, the network is trained by taking into account care-fully defined non-reference loss functions. This strategy al-lows output image quality to be implicitly evaluated, theresults of which would be reiterated for network learning.Third, our method is highly efficient and cost-effective.These advantages benefit from our zero-reference learn-ing framework, lightweight network structure, and effectivenon-reference loss functions.

3. MethodologyWe present the framework of Zero-DCE in Fig. 2. A

Deep Curve Estimation Network (DCE-Net) is devised toestimate a set of best-fitting Light-Enhancement curves(LE-curves) given an input image. The framework thenmaps all pixels of the input’s RGB channels by applyingthe curve parameters iteratively for obtaining the final en-hanced image. We next detail the key components in Zero-DCE, namely LE-curve, DCE-Net, and non-reference lossfunctions in the following sections.

3.1. Light-Enhancement Curve (LE-curve)

Inspired by the curves adjustment used in photo editingsoftware, we attempt to design a kind of curve that canmap a low-light image to its enhanced version automati-cally, where the self-adaptive curve parameters are solelydependent on the input image. There are three objectivesin the design of such a curve: 1) each pixel value of theenhanced image should be in the normalized range of [0,1]to avoid information loss induced by overflow truncation;

2) this curve should be monotonous to preserve the differ-ences (contrast) of neighboring pixels; and 3) the form ofthis curve should be as simple as possible and differentiablein the process of gradient backpropagation.

To achieve these three objectives, we design a quadraticcurve which can be expressed as:

LE(I(x);α) = I(x) + αI(x)(1− I(x)), (1)

where x denotes pixel coordinates, LE(I(x);α) is the en-hanced version of the given input I(x), α ∈ [−1, 1] is thetrainable curve parameter, which adjusts the magnitude ofLE-curve and also controls the exposure level. Each pixelis normalized to [0, 1] and all operations are pixel-wise.We separately apply the LE-curve to three RGB channelsinstead of solely on the illumination channel. The three-channel adjustment can better preserve the inherent colorand reduce the risk of over-saturation. We report more de-tails in the supplementary material.

An illustration of LE-curves with different adjustmentparameters α is shown in Fig. 2(b). It is clear that the LE-curve complies with the three aforementioned objectives.In addition, the LE-curve enables us to increase or decreasethe dynamic range of an input image. This capability isconducive to not only enhancing low-light regions but alsoremoving over-exposure artifacts.Higher-Order Curve. The LE-curve defined in Eq. (1) canbe applied iteratively to enable more versatile adjustment tocope with challenging low-light conditions. Specifically,

LEn(x) = LEn−1(x) + αnLEn−1(x)(1− LEn−1(x)),(2)

where n is the number of iteration, which controls the cur-vature. In this paper, we set the value of n to 8, which candeal with most cases satisfactory. Eq. (2) can be degraded toEq. (1) when n is equal to 1. Figure 2(c) provides an exam-ple showing high-order curves with differentα and n, whichhave more powerful adjustment capability (i.e., greater cur-vature) than the curves in Figure 2(b).Pixel-Wise Curve. A higher-order curve can adjust an im-age within a wider dynamic range. Nonetheless, it is stilla global adjustment since α is used for all pixels. A globalmapping tends to over-/under- enhance local regions. To ad-dress this problem, we formulate α as a pixel-wise param-eter, i.e., each pixel of the given input image has a corre-sponding curve with the best-fitting α to adjust its dynamicrange. Hence, Eq. (2) can be reformulated as:

LEn(x) = LEn−1(x)+An(x)LEn−1(x)(1−LEn−1(x)),(3)

whereA is a parameter map with the same size as the givenimage. Here, we assume that pixels in a local region havethe same intensity (also the same adjustment curves), andthus the neighboring pixels in the output result still pre-serve the monotonous relations. In this way, the pixel-wise

4323

Page 4: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

I

Enhanced ImageCurve Parameter Map

Deep Curve Estimation Network (DCE-Net)

Input, LE1 = LE(I; AR,G,B

1 )<latexit sha1_base64="+H0yOULSCcsOmsNdNx0s13nZPwo=">AAACD3icbVDLSsNAFJ3UV62vqks3g0WpUEpSBQURqlJU6KKKfUBTw2Q6bYdOHsxMhBLyB278FTcuFHHr1p1/46TNQqsHLhzOuZd777F9RoXU9S8tNTM7N7+QXswsLa+srmXXNxrCCzgmdewxj7dsJAijLqlLKhlp+Zwgx2akaQ/PY795T7ignnsrRz7pOKjv0h7FSCrJyu5WK5YBT2C1kr86Nh0kBxix8DS6C83wpnBRODOjyDL2rGxOL+pjwL/ESEgOJKhZ2U+z6+HAIa7EDAnRNnRfdkLEJcWMRBkzEMRHeIj6pK2oixwiOuH4nwjuKKULex5X5Uo4Vn9OhMgRYuTYqjO+WEx7sfif1w5k76gTUtcPJHHxZFEvYFB6MA4HdiknWLKRIghzqm6FeIA4wlJFmFEhGNMv/yWNUtHYL5auD3JlPYkjDbbANsgDAxyCMrgENVAHGDyAJ/ACXrVH7Vl7094nrSktmdkEv6B9fANlPZpV</latexit>

LE2 = LE(LE1; AR,G,B2 )

<latexit sha1_base64="EyGvn8XkHiYRRdp7jVFZL27UXJU=">AAACEnicbZDLSgMxFIYz9VbrrerSTbAILZQyUwUFEapSdNFFFXuBzjhk0rQNzVxIMkIZ5hnc+CpuXCji1pU738a0nYW2/hD4+M855JzfCRgVUte/tdTC4tLySno1s7a+sbmV3d5pCj/kmDSwz3zedpAgjHqkIalkpB1wglyHkZYzvBzXWw+EC+p7d3IUEMtFfY/2KEZSWXa2UKvaZXgGa9W8IuPUdJEcYMSi8/g+MqPb4lXxwoxju1ywszm9pE8E58FIIAcS1e3sl9n1cegST2KGhOgYeiCtCHFJMSNxxgwFCRAeoj7pKPSQS4QVTU6K4YFyurDnc/U8CSfu74kIuUKMXEd1jjcWs7Wx+V+tE8reiRVRLwgl8fD0o17IoPThOB/YpZxgyUYKEOZU7QrxAHGEpUoxo0IwZk+eh2a5ZByWyjdHuYqexJEGe2Af5IEBjkEFXIM6aAAMHsEzeAVv2pP2or1rH9PWlJbM7II/0j5/AD9wm00=</latexit>

LEn = LE(LEn1; AR,G,Bn )

<latexit sha1_base64="czAbCvE0/y9GnTbGNB6KsidP4CA=">AAACFnicbZDLSgMxFIYzXmu9VV26CRahQltmqqAgQlWKLrqoYi/QqUMmTdvQTGZIMkIZ5inc+CpuXCjiVtz5NqaXhbb+EPj4zznknN8NGJXKNL+NufmFxaXlxEpydW19YzO1tV2TfigwqWKf+aLhIkkY5aSqqGKkEQiCPJeRutu/HNbrD0RI6vM7NQhIy0NdTjsUI6UtJ5UrlxwOz2C5lNEU8ZwVn9oeUj2MWHQe30d2dJu9yl7YcezwAyeVNvPmSHAWrAmkwUQVJ/Vlt30ceoQrzJCUTcsMVCtCQlHMSJy0Q0kChPuoS5oaOfKIbEWjs2K4r5027PhCP67gyP09ESFPyoHn6s7hxnK6NjT/qzVD1TlpRZQHoSIcjz/qhAwqHw4zgm0qCFZsoAFhQfWuEPeQQFjpJJM6BGv65FmoFfLWYb5wc5QumpM4EmAX7IEMsMAxKIJrUAFVgMEjeAav4M14Ml6Md+Nj3DpnTGZ2wB8Znz8y6p2A</latexit>

AR,G,Bn

<latexit sha1_base64="B6yDr0fvwKWcu1z0ZaETWvBM02E=">AAACAnicbVBNS8NAEJ3Ur1q/qp7ES7AIHkpJqqDHqgc9VrEf0MSy2W7bpZtN2N0IJQQv/hUvHhTx6q/w5r9x0+agrQ8GHu/NMDPPCxmVyrK+jdzC4tLySn61sLa+sblV3N5pyiASmDRwwALR9pAkjHLSUFQx0g4FQb7HSMsbXaZ+64EISQN+p8YhcX004LRPMVJa6hb3HB+pIUYsPk/uYye+LV+VL5wk6WqvZFWsCcx5YmekBBnq3eKX0wtw5BOuMENSdmwrVG6MhKKYkaTgRJKECI/QgHQ05cgn0o0nLyTmoVZ6Zj8QurgyJ+rviRj5Uo59T3emB8tZLxX/8zqR6p+5MeVhpAjH00X9iJkqMNM8zB4VBCs21gRhQfWtJh4igbDSqRV0CPbsy/OkWa3Yx5XqzUmpZmVx5GEfDuAIbDiFGlxDHRqA4RGe4RXejCfjxXg3PqatOSOb2YU/MD5/AOc0lww=</latexit>

AR,G,B1<latexit sha1_base64="F3tVJcaYu9Kt4WOxVxPugel5988=">AAACAnicbVDLSsNAFJ3UV62vqCtxM1gEF6UkVdBl1YUuq9gHNDFMptN26GQSZiZCCcGNv+LGhSJu/Qp3/o2TNgttPXDhcM693HuPHzEqlWV9G4WFxaXlleJqaW19Y3PL3N5pyTAWmDRxyELR8ZEkjHLSVFQx0okEQYHPSNsfXWZ++4EISUN+p8YRcQM04LRPMVJa8sw9J0BqiBFLztP7xEluK1eVCydNPdszy1bVmgDOEzsnZZCj4ZlfTi/EcUC4wgxJ2bWtSLkJEopiRtKSE0sSITxCA9LVlKOASDeZvJDCQ630YD8UuriCE/X3RIICKceBrzuzg+Wsl4n/ed1Y9c/chPIoVoTj6aJ+zKAKYZYH7FFBsGJjTRAWVN8K8RAJhJVOraRDsGdfnietWtU+rtZuTsp1K4+jCPbBATgCNjgFdXANGqAJMHgEz+AVvBlPxovxbnxMWwtGPrML/sD4/AGKwJbP</latexit>

AR,G,B2<latexit sha1_base64="jRKNyUsq/nSHwEAPmF/2wCST14c=">AAACAnicbVDLSsNAFJ3UV62vqCtxM1gEF6UkVdBl1YUuq9gHNDFMptN26GQSZiZCCcGNv+LGhSJu/Qp3/o2TNgttPXDhcM693HuPHzEqlWV9G4WFxaXlleJqaW19Y3PL3N5pyTAWmDRxyELR8ZEkjHLSVFQx0okEQYHPSNsfXWZ++4EISUN+p8YRcQM04LRPMVJa8sw9J0BqiBFLztP7xEluK1eVCydNvZpnlq2qNQGcJ3ZOyiBHwzO/nF6I44BwhRmSsmtbkXITJBTFjKQlJ5YkQniEBqSrKUcBkW4yeSGFh1rpwX4odHEFJ+rviQQFUo4DX3dmB8tZLxP/87qx6p+5CeVRrAjH00X9mEEVwiwP2KOCYMXGmiAsqL4V4iESCCudWkmHYM++PE9atap9XK3dnJTrVh5HEeyDA3AEbHAK6uAaNEATYPAInsEreDOejBfj3fiYthaMfGYX/IHx+QOMRJbQ</latexit> …

I<latexit sha1_base64="exkWYHz5kAdkFbhPQ/1u6ncFZak=">AAAB6HicbVBNS8NAEJ3Ur1q/qh69LBbBU0mqoMeCF721YGuhDWWznbRrN5uwuxFK6C/w4kERr/4kb/4bt20O2vpg4PHeDDPzgkRwbVz32ymsrW9sbhW3Szu7e/sH5cOjto5TxbDFYhGrTkA1Ci6xZbgR2EkU0igQ+BCMb2b+wxMqzWN5byYJ+hEdSh5yRo2Vmnf9csWtunOQVeLlpAI5Gv3yV28QszRCaZigWnc9NzF+RpXhTOC01Es1JpSN6RC7lkoaofaz+aFTcmaVAQljZUsaMld/T2Q00noSBbYzomakl72Z+J/XTU147WdcJqlByRaLwlQQE5PZ12TAFTIjJpZQpri9lbARVZQZm03JhuAtv7xK2rWqd1GtNS8rdTePowgncArn4MEV1OEWGtACBgjP8ApvzqPz4rw7H4vWgpPPHMMfOJ8/muGMvw==</latexit>

(a)

(b)

(c)

Figure 2. (a) The framework of Zero-DCE. A DCE-Net is devised to estimate a set of best-fitting Light-Enhancement curves (LE-curves)to iteratively enhance a given input image. (b, c) LE-curves with different adjustment parameters α and numbers of iteration n. In (c), α1,α2, and α3 are equal to -1 while n is equal to 4. In each subfigure, the horizontal axis represents the input pixel values while the verticalaxis represents the output pixel values.

(a) Input (b) ARn (c) AG

n (d) ABn (e) Result

Figure 3. An example of the pixel-wise curve parameter maps. For visualization, we average the curve parameter maps of all iterations(n = 8) and normalize the values to the range of [0, 1]. AR

n , AGn , and AB

n represent the averaged best-fitting curve parameter maps of R,G, and B channels, respectively. The maps in (b), (c), and (d) are represented by heatmaps.

higher-order curves also comply with the above-mentionedthree objectives.

We present an example of the estimated curve parame-ter maps of three channels in Fig. 3. As shown, the best-fitting parameter maps of different channels have similaradjustment tendency but different values, indicating the rel-evance and difference among the three channels of a low-light image. The curve parameter map accurately indicatesthe brightness of different regions (e.g., the two glitters onthe wall). With the fitting maps, the enhanced version imagecan be directly obtained by pixel-wise curve mapping. Asshown in Fig. 3(e), the enhanced version reveals the contentin dark regions and preserves the bright regions.

3.2. DCE-Net

To learn the mapping between an input image and itsbest-fitting curve parameter maps, we propose a DeepCurve Estimation Network (DCE-Net). The input to theDCE-Net is a low-light image while the outputs are a set ofpixel-wise curve parameter maps for corresponding higher-order curves. We employ a plain CNN of seven convolu-

tional layers with symmetrical concatenation. Each layerconsists of 32 convolutional kernels of size 3×3 and stride1 followed by the ReLU activation function. We discard thedown-sampling and batch normalization layers that breakthe relations of neighboring pixels. The last convolutionallayer is followed by the Tanh activation function, which pro-duces 24 parameter maps for 8 iterations (n = 8), whereeach iteration requires three curve parameter maps for thethree channels. The detailed architecture of DCE-Net isprovided in the supplementary material. It is noteworthythat DCE-Net only has 79,416 trainable parameters and5.21G Flops for an input image of size 256 ×256 ×3. Itis therefore lightweight and can be used in computationalresource-limited devices, such as mobile platforms.

3.3. Non-Reference Loss Functions

To enable zero-reference learning in DCE-Net, we pro-pose a set of differentiable non-reference losses that allowus to evaluate the quality of enhanced images. The follow-ing four types of losses are adopted to train our DCE-Net.Spatial Consistency Loss. The spatial consistency loss

4324

Page 5: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

Lspa encourages spatial coherence of the enhanced imagethrough preserving the difference of neighboring regionsbetween the input image and its enhanced version:

Lspa =1

K

K∑i=1

∑j∈Ω(i)

(‖(Yi − Yj)‖ − ‖(Ii − Ij)‖)2, (4)

whereK is the number of local region i, and Ω(i) is the fourneighboring regions (top, down, left, right) centered at theregion i. We denote Y and I as the average intensity valueof the local region in the enhanced version and input image,respectively. We empirically set the size of the local regionto 4×4. This loss is stable given other region sizes.Exposure Control Loss. To restrain under-/over-exposedregions, we design an exposure control loss Lexp to con-trol the exposure level. The exposure control loss measuresthe distance between the average intensity value of a localregion to the well-exposedness level E. We follow existingpractices [19, 20] to setE as the gray level in the RGB colorspace. We set E to 0.6 in our experiments although we donot find much performance difference by setting E within[0.4, 0.7]. The loss Lexp can be expressed as:

Lexp =1

M

∑M

k=1‖Yk − E‖, (5)

whereM represents the number of nonoverlapping local re-gions of size 16×16, Y is the average intensity value of alocal region in the enhanced image.Color Constancy Loss. Following Gray-World color con-stancy hypothesis [2] that color in each sensor channel av-erages to gray over the entire image, we design a color con-stancy loss to correct the potential color deviations in theenhanced image and also build the relations among the threeadjusted channels. The color constancy loss Lcol can be ex-pressed as:

Lcol =∑∀(p,q)∈ε

(Jp − Jq)2, ε = R,G,B, (6)

where Jp denotes the average intensity value of p channelin the enhanced image, (p,q) represents a pair of channels.Illumination Smoothness Loss. To preserve the mono-tonicity relations between neighboring pixels, we add anillumination smoothness loss to each curve parameter mapA. The illumination smoothness loss LtvA is defined as:

LtvA =1

N

N∑n=1

∑c∈ε

(∇xAcn +∇yAc

n)2, ε = R,G,B,

(7)where N is the number of iteration, ∇x and ∇y representthe horizontal and vertical gradient operations, respectively.Total Loss. The total loss can be expressed as:

Ltotal = Lspa + Lexp +WcolLcol +WtvALtvA , (8)

where Wcol and WtvA are the weights of the losses.

4. ExperimentsImplementation Details. CNN-based models usually useself-captured paired data for network training [27, 5, 24]while GAN-based models elaborately select unpaired data[10]. To bring the capability of wide dynamic range adjust-ment into full play, we incorporate both low-light and over-exposed images into our training set. To this end, we em-ploy 360 multi-exposure sequences from the Part1 of SICEdataset [4] to train the proposed DCE-Net. The dataset isalso used as a part of the training data in EnlightenGAN[10]. We randomly split 3,022 images of different exposurelevels in the Part1 subset [4] into two parts (2,422 imagesfor training and the rest for validation). We resize the train-ing images to the size of 512×512.

We implement our framework with PyTorch on anNVIDIA 2080Ti GPU. A batch size of 8 is applied. The fil-ter weights of each layer are initialized with standard zeromean and 0.02 standard deviation Gaussian function. Biasis initialized as a constant. We use ADAM optimizer withdefault parameters and fixed learning rate 1e−4 for our net-work optimization.

4.1. Ablation Study

We perform several ablation studies to demonstrate theeffectiveness of each component of Zero-DCE as follows.More qualitative and quantitative comparisons can be foundin the supplementary material.Contribution of Each Loss. We present the results of Zero-DCE trained by various combinations of losses in Fig. 4.The result without spatial consistency loss Lspa has rela-tively lower contrast (e.g., the cloud regions) than the fullresult. This shows the importance of Lspa in preservingthe difference of neighboring regions between the input andthe enhanced image. Removing the exposure control lossLexp fails to recover the low-light region. Severe colorcasts emerge when the color constancy loss Lcol is dis-carded. This variant ignores the relations among three chan-nels when curve mapping is applied. Finally, removing theillumination smoothness loss LtvA hampers the correlationsbetween neighboring regions leading to obvious artifacts.Effect of Parameter Settings. We evaluate the effect ofparameters in Zero-DCE, consisting of the depth and widthof the DCE-Net and the number of iterations. A visualexample is presented in Fig. 5. In Fig. 5(b), with justthree convolutional layers, Zero-DCE3−32−8 can alreadyproduce satisfactory results, suggesting the effectiveness ofzero-reference learning. The Zero-DCE7−32−8 and Zero-DCE7−32−16 produce most visually pleasing results withnatural exposure and proper contrast. By reducing the num-ber of iterations to 1, an obvious decrease in performance isobserved on Zero-DCE7−32−1 as shown in Fig. 5(d). Thisis because the curve with only single iteration has limitedadjustment capability. This suggests the need for higher-

4325

Page 6: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

(a) Input (b) Zero-DCE (c) w/o Lspa (d) w/o Lexp (e) w/o Lcol (f) w/o LtvA

Figure 4. Ablation study of the contribution of each loss (spatial consistency loss Lspa, exposure control loss Lexp, color constancy lossLcol, illumination smoothness loss LtvA ).

(a) Input (b) 3-32-8 (c) 7-16-8

(d) 7-32-1 (e) 7-32-8 (f) 7-32-16

Figure 5. Ablation study of the effect of parameter settings. l-f -nrepresents the proposed Zero-DCE with l convolutional layers, ffeature maps of each layer (except the last layer), and n iterations.

order curves in our method. We choose Zero-DCE7−32−8

as the final model based given its good trade-off betweenefficiency and restoration performance.Impact of Training Data. To test the impact of trainingdata, we retrain the Zero-DCE on different datasets: 1) only900 low-light images out of 2,422 images in the originaltraining set (Zero-DCELow), 2) 9,000 unlabeled low-lightimages provided in the DARK FACE dataset [30] (Zero-DCELargeL), and 3) 4800 multi-exposure images from thedata augmented combination of Part1 and Part2 subsets inthe SICE dataset [4] (Zero-DCELargeLH ). As shown inFig. 6(c) and (d), after removing the over-exposed trainingdata, Zero-DCE tends to over-enhance the well-lit regions(e.g., the face), in spite of using more low-light images, (i.e.,Zero-DCELargeL). Such results indicate the rationality andnecessity of the usage of multi-exposure training data inthe training process of our network. In addition, the Zero-DCE can better recover the dark regions when more multi-exposure training data are used (i.e., Zero-DCELargeLH ),as shown in Fig. 6(e). For a fair comparison with otherdeep learning-based methods, we use a comparable amountof training data with them although more training data canbring better visual performance to our approach.

4.2. Benchmark Evaluations

We compare Zero-DCE with several state-of-the-artmethods: three conventional methods (SRIE [7], LIME [8],

Li et al. [15]), two CNN-based methods (RetinexNet [27],Wang et al. [24] ), and one GAN-based method (Enlighten-GAN [10]). The results are reproduced by using publiclyavailable source codes with recommended parameters.

We perform qualitative and quantitative experimentson standard image sets from previous works includingNPE [25] (84 images), LIME [8] (10 images), MEF [18](17 images), DICM [12] (64 images), and VV‡ (24 im-ages). Besides, we quantitatively validate our method onthe Part2 subset of SICE dataset [4], which consists of 229multi-exposure sequences and the corresponding referenceimage for each multi-exposure sequence. For a fair com-parison, we only use the low-light images of Part2 sub-set [4] for testing, since baseline methods cannot handleover-exposed images well. Specifically, we choose the firstthree (resp. four) low-light images if there are seven (resp.nine) images in a multi-exposure sequence and resize allimages to a size of 1200×900×3. Finally, we obtain 767paired low/normal light images. We discard the low/normallight image dataset mentioned in [30], because the trainingdatasets of RetinexNet [27] and EnlightenGAN [10] con-sist of some images from this dataset. Note that the lat-est paired training and testing dataset constructed in [24]are not publicly available. We did not use the MIT-AdobeFiveK dataset [3] as it is not primarily designed for under-exposed photos enhancement. A discussion is provided in[24].

4.2.1 Visual and Perceptual Comparisons

We present the visual comparisons on typical low-light im-ages in Fig. 7. For challenging back-lit regions (e.g., theface in Fig. 7(a)), Zero-DCE yields natural exposure andclear details while SRIE [7], LIME [8], Wang et al. [24],and EnlightenGAN [10] cannot recover the face clearly.RetinexNet [27] produces over-exposed artifacts. In thesecond example featuring an indoor scene, our method en-hances the dark regions and preserves the color of the in-put image simultaneously. The result is visually pleasingwithout obvious noise and color casts. In contrast, the Li etal. [15] over-smoothes the details while other baselines am-

‡https://sites.google.com/site/vonikakis/datasets

4326

Page 7: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

(a) Input (b) Zero-DCE (c) Zero-DCELow (d) Zero-DCELargeL (e) Zero-DCELargeLH

Figure 6. Ablation study on the impact of training data.

(a) Inputs (b) SRIE [7] (c) LIME [8] (d) Li et al. [15]

(e) RetinexNet [27] (f) Wang et al. [24] (g) EnlightenGAN [10] (h) Zero-DCE

Figure 7. Visual comparisons on typical low-light images. Red boxes indicate the obvious differences.

plify noise and even produce color deviation (e.g., the colorof wall).

We perform a user study to quantify the subjective visualquality of various methods. We process low-light imagesfrom the image sets (NPE, LIME, MEF, DICM, VV) by dif-ferent methods. For each enhanced result, we display it ona screen and provide the input image as a reference. A to-tal of 15 human subjects are invited to independently scorethe visual quality of the enhanced image. These subjectsare trained by observing the results from 1) whether the re-sults contain over-/under-exposed artifacts or over-/under-enhanced regions; 2) whether the results introduce colordeviation; and 3) whether the results have unnatural tex-ture and obvious noise. The scores of visual quality rangefrom 1 to 5 (worst to best quality). The average subjectivescores for each image set are reported in Table 1. As sum-marized in Table 1, Zero-DCE achieves the highest averageUser Study (US) score for a total of 202 testing images from

the above-mentioned image sets. For the MEF, DICM, andVV sets, our results are most favored by the subjects. Inaddition to the US score, we also employ a non-referenceperceptual index (PI) [1, 17, 21] to evaluate the perceptualquality of our results. The PI metric is originally used tomeasure perceptual quality in image super-resolution. Ithas also been used to assess the performance of other im-age restoration tasks, such as image dehazing [22]. A lowerPI value indicates better perceptual quality. The PI valuesare reported in Table 1 too. Similar to the user study, theproposed Zero-DCE is superior to other competing meth-ods in terms of the average PI values. It obtains the bestperceptual quality on LIME, MEF, and DICM sets.

4.2.2 Quantitative Comparisons

For full-reference image quality assessment, we employ thePeak Signal-to-Noise Ratio (PSNR,dB), Structural Similar-ity (SSIM) [26], and Mean Absolute Error (MAE) met-

4327

Page 8: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

Table 1. User study (US)↑/Perceptual index (PI)↓ scores on the image sets (NPE, LIME, MEF, DICM, VV). Higher US score indicatesbetter human subjective visual quality while lower PI value indicates better perceptual quality. The best result is in red whereas the secondbest one is in blue under each case.

Method NPE LIME MEF DICM VV AverageSRIE [7] 3.65/2.79 3.50/2.76 3.22/2.61 3.42/3.17 2.80/3.37 3.32/2.94LIME [8] 3.78/3.05 3.95/3.00 3.71/2.78 3.31/3.35 3.21/3.03 3.59/3.04

Li et al. [15] 3.80/3.09 3.78/3.02 2.93/3.61 3.47/3.43 2.87/3.37 3.37/3.72RetinexNet [27] 3.30/3.18 2.32/3.08 2.80/2.86 2.88/3.24 1.96/2.95 2.58/3.06Wang et al. [24] 3.83/2.83 3.82/2.90 3.13/2.72 3.44/3.20 2.95/3.42 3.43/3.01

EnlightenGAN [10] 3.90/2.96 3.84/2.83 3.75/2.45 3.50/3.13 3.17/4.71 3.63/3.22Zero-DCE 3.81/2.84 3.80/2.76 4.13/2.43 3.52/3.04 3.24/3.33 3.70/2.88

Table 2. Quantitative comparisons in terms of full-reference imagequality assessment metrics. The best result is in red whereas thesecond best one is in blue under each case.

Method PSNR↑ SSIM↑ MAE↓SRIE [7] 14.41 0.54 127.08LIME [8] 16.17 0.57 108.12

Li et al. [15] 15.19 0.54 114.21RetinexNet [27] 15.99 0.53 104.81Wang et al. [24] 13.52 0.49 142.01

EnlightenGAN [10] 16.21 0.59 102.78Zero-DCE 16.57 0.59 98.78

Table 3. Runtime (RT) comparisons (in second). The best result isin red whereas the second best one is in blue.

Method RT PlatformSRIE [7] 12.1865 MATLAB (CPU)LIME [8] 0.4914 MATLAB (CPU)

Li et al. [15] 90.7859 MATLAB (CPU)RetinexNet [27] 0.1200 TensorFlow (GPU)Wang et al. [24] 0.0210 TensorFlow (GPU)

EnlightenGAN [10] 0.0078 PyTorch (GPU)Zero-DCE 0.0025 PyTorch (GPU)

rics to quantitatively compare the performance of differ-ent methods on the Part2 subset [4]. As shown in Ta-ble 2, the proposed Zero-DCE achieves the best values un-der all cases, despite that it does not use any paired orunpaired training data. Zero-DCE is also computationallyefficient, benefited from the simple curve mapping formand lightweight network structure. Table 3 shows the run-time§ of different methods averaged on 32 images of size1200×900×3. For conventional methods, only the codes ofCPU version are available.

4.2.3 Face Detection in the Dark

We investigate the performance of low-light image en-hancement methods on the face detection task under low-

§Runtime is measured on a PC with an Nvidia GTX 2080Ti GPU andIntel I7 6700 CPU, except for Wang et al. [24], which has to run on GTX1080Ti GPU.

light conditions. Specifically, we use the latest DARKFACE dataset [30] that composes of 10,000 images takenin the dark. Since the bounding boxes of test set are notpublicly available, we perform evaluation on the trainingand validation sets, which consists of 6,000 images. Astate-of-the-art deep face detector, Dual Shot Face Detec-tor (DSFD) [14], trained on WIDER FACE dataset [28], isused as the baseline model. We feed the results of differentlow-light image enhancement methods to the DSFD [14]and depict the precision-recall (P-R) curves in Fig. 8. Be-sides, we also compare the average precision (AP) by usingthe evaluation tool¶ provided in DARK FACE dataset [30].

Raw Detection Enhanced Detection

Figure 8. The performance of face detection in the dark. P-Rcurves, the AP, and two examples of face detection before andafter enhanced by our Zero-DCE.

As shown in Fig. 8, after image enhancement, the preci-sion of DSFD [14] increases considerably compared to thatusing raw images without enhancement. Among differentmethods, RetinexNet [27] and Zero-DCE perform the best.Both methods are comparable but Zero-DCE performs bet-ter in the high recall area. Observing the examples, our¶https://github.com/Ir1d/DARKFACE_eval_tools

4328

Page 9: Zero-Reference Deep Curve Estimation for Low-Light Image ... · DCE, namely LE-curve, DCE-Net, and non-reference loss functions in the following sections. 3.1. Light-Enhancement Curve

Zero-DCE lightens up the faces in the extremely dark re-gions and preserves the well-lit regions, thus improves theperformance of face detector in the dark.

5. ConclusionWe proposed a novel deep network for low-light image

enhancement. It can be trained end-to-end with zero refer-ence images. This is achieved by formulating the low-lightimage enhancement task as an image-specific curve esti-mation problem, and devising a set of differentiable non-reference losses. Experiments convincingly demonstratethe superiority of our method against existing low-light im-age enhancement methods.

References[1] Yochai Blau and Tomer Michaeli. The perception-distortion

tradeoff. In CVPR, 2018.[2] Gershon Buchsbaum. A spatial processor model for object

colour perception. J. Franklin Institute, 310(1):1–26, 1980.[3] Vladimir Bychkovsky, Sylvain Paris, Eric Chan, and Fredo

Durand. Learning photographic global tonal adjustment witha database of input/output image pairs. In CVPR, 2011.

[4] Jianrui Cai, Shuhang Gu, and Lei Zhang. Learning a deepsingle image contrast enhancer from multi-exposure image.IEEE Transactions on Image Processing, 27(4):2049–2026,2018.

[5] Chen Chen, Qifeng Chen, Jia Xu, and Koltun Vladlen.Learning to see in the dark. In CVPR, 2018.

[6] Dinu Coltuc, Philippe Bolon, and Jean-Marc Chassery. Ex-act histogram specification. IEEE Transactions on ImageProcessing, 15(5):1143–1152, 2006.

[7] Xueyang Fu, Delu Zeng, Yue Huang, Xiao-Ping Zhang, andXinghao Ding. A weighted variational model for simultane-ous reflectance and illumination estimation. In CVPR, 2016.

[8] Xiaojie Guo, Yu Li, and Haibin Ling. Lime: Low-light im-age enhancement via illumination map estimation. IEEETransactions on Image Processing, 26(2):982–993, 2017.

[9] Haidi Ibrahim and Nicholas Sia Pik Kong. Brightness pre-serving dynamic histogram equalization for image contrastenhancement. IEEE Transactions on Consumer Electronics,53(4):1752–1758, 2007.

[10] Yifan Jiang, Xinyu Gong, Ding Liu, Yu Cheng, Chen Fang,Xiaohui Shen, Jianchao Yang, Pan Zhou, and ZhangyangWang. EnlightenGAN: Deep light enhancement withoutpaired supervision. In CVPR, 2019.

[11] Edwin H Land. The retinex theory of color vision. ScientificAmerican, 237(6):108–128, 1977.

[12] Chulwoo Lee, Chul Lee, and Chang-Su Kim. Contrast en-hancement based on layered difference representation. InICIP, 2012.

[13] Chulwoo Lee, Chul Lee, and Chang-Su Kim. Contrastenhancement based on layered difference representation of2d histograms. IEEE Transactions on Image Processing,22(12):5372–5384, 2013.

[14] Jian Li, Yabiao Wang, Changan Wang, Ying Tai, JianjunQian, Jian Yang, Chengjie Wang, Jilin Li, and FeiyuenHuang. Dsfd: Dual shot face detector. In CVPR, 2019.

[15] Mading Li, Jiaying Liu, Wenhan Yang, Xiaoyan Sun, andZongming Guo. Structure-revealing low-light image en-hancement via robust retinex model. IEEE Transactions onImage Processing, 27(6):2828–2841, 2018.

[16] Kin Gwn Lore, Adedotun Akintayo, and Soumik Sarkar. Ll-net: A deep autoencoder approach to natural low-light imageenhancement. Pattern Recognition, 61:650–662, 2017.

[17] Chao Ma, Chih-Yuan Yang, Xiaokang Yang, and Ming-Hsuan Yang. Learning a no-reference quality metric forsingle-image super-resolution. Computer Vision and ImageUnderstanding, 158:1–16, 2017.

[18] Kede Ma, Kai Zeng, and Zhou Wang. Perceptual qualityassessment for multi-exposure image fusion. IEEE Transac-tions on Image Processing, 24(11):3345–3356, 2015.

[19] Tom Mertens, Jan Kautz, and Frank Van Reeth. Exposurefusion. In PCCGA, 2007.

[20] Tom Mertens, Jan Kautz, and Frank Van Reeth. Exposurefusion: A simple and practrical alterrnative to high dynamicrange photography. Computer Graphics Forum, 28(1):161–171, 2009.

[21] Anish Mittal, Rajiv Soundararajan, and Alan C. Bovik. Mak-ing a “completely blind” image quality analyzer. IEEE Sig-nal Processing Letters, 20(3):209–212, 2013.

[22] Yanyun Qu, Yizi Chen, Jingying Huang, and Yuan Xie. En-hanced pix2pix dehazing network. In CVPR, 2019.

[23] J Alex Stark. Adaptive image contrast enhancement usinggeneralizations of histogram equalization. IEEE Transac-tions on Image Processing, 9(5):889–896, 2000.

[24] Ruixing Wang, Qing Zhang, Chi-Wing Fu, Xiaoyong Shen,Wei-Shi Zheng, and Jiaya Jia. Underexposed photo enhance-ment using deep illumination estimation. In CVPR, 2019.

[25] Shuhang Wang, Jin Zheng, Hai-Miao Hu, and Bo Li. Nat-uralness preserved enhancement algorithm for non-uniformillumination images. IEEE Transactions on Image Process-ing, 22(9):3538–3548, 2013.

[26] Zhou Wang, Alan C. Bovik, Hamid R. Sheikh, and Eero P.Simoncelli. Image quality assessment: From error visibilityto structural similarity. IEEE Transactions on Image Pro-cessing, 13(4):600–612, 2004.

[27] Chen Wei, Wenjing Wang, Wenhan Yang, and Jiaying Liu.Deep retinex decomposition for low-light enhancement. InBMVC, 2018.

[28] Shuo Yang, Ping Luo, Chen-Change Loy, and Xiaoou Tang.Wider face: A face detection benchmark. In CVPR, 2016.

[29] Lu Yuan and Jian Sun. Automatic exposure correction ofconsumer photographs. In ECCV, 2012.

[30] Ye Yuan, Wenhan Yang, Wenqi Ren, Jiaying Liu, Walter JScheirer, and Wang Zhangyang. Ug+ track 2: A collec-tive benchmark effort for evaluating and advancing imageunderstanding in poor visibility environments, 2019. arXivarXiv:1904.04474.

[31] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei AEfros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In ICCV, 2017.

4329