arxiv:2108.08826v1 [cs.cv] 19 aug 2021

20
Towards Vivid and Diverse Image Colorization with Generative Color Prior Yanze Wu, Xintao Wang, Yu Li, Honglun Zhang, Xun Zhao, Ying Shan Applied Research Center (ARC), Tencent PCG {yanzewu,xintaowang,ianyli,honlanzhang,emmaxunzhao,yingsshan}@tencent.com Inputs Results Diverse colorization results Controllable and smooth transitions Figure 1: By leveraging the generative color prior and the delicate designs, our method is capable of generating vivid and diverse colorization results without relying on external exemplars. Besides, our method could attain controllable and smooth transitions by walking through the GAN latent space. Abstract Colorization has attracted increasing interest in recent years. Classic reference-based methods usually rely on external color images for plausible results. A large im- age database or online search engine is inevitably required for retrieving such exemplars. Recent deep-learning-based methods could automatically colorize images at a low cost. However, unsatisfactory artifacts and incoherent colors are always accompanied. In this work, we aim at recovering vivid colors by leveraging the rich and diverse color priors encapsulated in a pretrained Generative Adversarial Net- works (GAN). Specifically, we first “retrieve” matched fea- tures (similar to exemplars) via a GAN encoder and then in- corporate these features into the colorization process with feature modulations. Thanks to the powerful generative color prior and delicate designs, our method could pro- duce vivid colors with a single forward pass. Moreover, it is highly convenient to obtain diverse results by modify- ing GAN latent codes. Our method also inherits the merit of interpretable controls of GANs and could attain control- lable and smooth transitions by walking through GAN latent space. Extensive experiments and user studies demonstrate that our method achieves superior performance than previ- ous works. 1. Introduction Colorization, the task of restoring colors from black- and-white photos, has wide applications in various fields, such as photography technologies, advertising or film in- arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Upload: others

Post on 03-Nov-2021

15 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Towards Vivid and Diverse Image Colorization with Generative Color Prior

Yanze Wu, Xintao Wang, Yu Li, Honglun Zhang, Xun Zhao, Ying ShanApplied Research Center (ARC), Tencent PCG

{yanzewu,xintaowang,ianyli,honlanzhang,emmaxunzhao,yingsshan}@tencent.com

Inpu

tsR

esul

ts

Diverse colorization results

Controllable and smooth transitions

Figure 1: By leveraging the generative color prior and the delicate designs, our method is capable of generating vivid anddiverse colorization results without relying on external exemplars. Besides, our method could attain controllable and smoothtransitions by walking through the GAN latent space.

Abstract

Colorization has attracted increasing interest in recentyears. Classic reference-based methods usually rely onexternal color images for plausible results. A large im-age database or online search engine is inevitably requiredfor retrieving such exemplars. Recent deep-learning-basedmethods could automatically colorize images at a low cost.However, unsatisfactory artifacts and incoherent colors arealways accompanied. In this work, we aim at recoveringvivid colors by leveraging the rich and diverse color priorsencapsulated in a pretrained Generative Adversarial Net-works (GAN). Specifically, we first “retrieve” matched fea-tures (similar to exemplars) via a GAN encoder and then in-corporate these features into the colorization process with

feature modulations. Thanks to the powerful generativecolor prior and delicate designs, our method could pro-duce vivid colors with a single forward pass. Moreover,it is highly convenient to obtain diverse results by modify-ing GAN latent codes. Our method also inherits the meritof interpretable controls of GANs and could attain control-lable and smooth transitions by walking through GAN latentspace. Extensive experiments and user studies demonstratethat our method achieves superior performance than previ-ous works.

1. IntroductionColorization, the task of restoring colors from black-

and-white photos, has wide applications in various fields,such as photography technologies, advertising or film in-

arX

iv:2

108.

0882

6v1

[cs

.CV

] 1

9 A

ug 2

021

Page 2: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

dustry [21, 48]. As colorization requires estimating missingcolor channels from only one grayscale value, it is inher-ently an ill-posed problem. Moreover, the plausible solu-tions of colorization are not unique (e.g., cars in black, blue,or red are all feasible results). Due to the uncertainty and di-versity nature of colorization, it remains a challenging task.

Classic reference-based methods (e.g. [17, 35, 53]) re-quire additional example color images as guidance. Thesemethods attempt to match the relevant contents between ex-emplars and input gray images, and then transfer the colorstatistics from the reference to the gray one. The qualityof generated colors (e.g., realness and vividness) is stronglydependent on the reference images. However, retrieving de-sirable reference images requires significant user efforts. Arecommendation system for this can be a solution, but it isstill challenging to design such a retrieval process. A large-scale color image database or online search engine is in-evitably required in the system.

Recently, convolutional neural network (CNN) basedcolorization methods [7, 11, 23] have been proposed to fa-cilitate the colorization task in an automatic fashion. Thesemethods learn to discover the semantics [31, 58] and thendirectly predict the colorization results. Though remarkableachievements in visual qualities are achieved, unsatisfactoryartifacts and incoherent colors are always accompanied.

In this work, we attempt to exploit the merits of bothreference-based and CNN-based methods, i.e., achievingrealistic and vivid colorization results on par with reference-based methods, while keeping them automatic at a low cost.Inspired by recent success in Generative Adversarial Net-work (GAN) [3, 27], our core idea is to leverage the mostrelevant image in the learned GAN distribution as the exem-plar image. The rich and diverse color information encapsu-lated in pretrained GAN models, i.e., generative color prior,allows us to circumvent the explicit example retrieval stepand integrate it into the colorization pipeline as a unifiedframework. Specifically, we first ‘retrieve’ matched fea-tures (similar to exemplars) via a GAN encoder and thenincorporate these features into the colorization process withfeature modulations. In addition, our method is capableof achieving diverse colorization from different samples inthe GAN distribution or by modifying GAN latent codes.Thanks to the interpretable controls of GANs, our methodcould also attain controllable and smooth transitions bywalking through GAN latent space.

We summarize the contributions as follows. (1) Wedevelop a unified framework to leverage rich and diversegenerative color prior for automatic colorization. (2) Ourmethod allows us to obtain diverse outputs. Controllableand smooth transitions can be achieved by manipulating thecode in GAN latent space. (3) Experiments show that ourmethod is capable of generating more vivid and diverse col-orization results than previous works.

2. Related WorkUser assisted colorization. Early colorization methods aremostly interactive and require users to draw color strokes onthe gray image to guide the colorization. The propagationfrom local hints to all pixels is usually formulated in a con-strained optimization manner [33] that tries to assign twopixels with the same color if they are adjacent and similarunder similarity measures. Attempts in this direction focuson finding efficient propagation [55] with different similar-ity metrics by hand-crafting [43] or learning [12], buildinglong-range relations[52], or using edge detection to reducebleeding [19]. While these methods can generate appeal-ing results with careful interactions, it demands intensiveuser efforts. The work of [59] alleviates this issue by usingsparse color points, and introduces a neural network to col-orize from these hints. Some methods also propose to useglobal hints like color palettes [2, 5] instead of dense colorpoints as constraints.Reference-based methods try to transfer the color statis-tics from the reference to the gray image using correspon-dences between the two based on low-level similarity mea-sures [34], semantic features [6], or super-pixels [8, 15].Recent works [17, 35, 53] adopt deep neural network to im-prove the spatial correspondence and colorization results.Though these methods obtain remarkable results when suit-able references are available, the procedure of finding ref-erences is time-consuming and challenging for automaticretrieval system [8].Automatic colorization methods are generally learningbased [7, 11, 23, 30]. Pre-trained networks for classifica-tion are used for better semantic representation [31]. Twobranch dual-task structures are also proposed in [22, 48, 60,61] that jointly learn the pixel embedding and local (seman-tic maps) or global (class labels) information. The recentwork [47] investigates the instance-level features to modelthe appearance variations of objects. To improve the di-versity of generated colors, Zhang et al. [58] uses a class-rebalancing scheme, while some methods propose to predictper-pixel color distributions [10, 39, 44] instead of a singlecolor. Yet, their results typically suffer from visual artifactslike unnatural and incoherent colors.Generative priors of pretrained GANs [3, 26, 27, 28] ispreviously exploited by GAN inversion [1, 13, 38, 41, 62,64], which aims to find the closest latent codes given an in-put image. In colorization, they first ‘invert’ the grayscaleimage back to a latent code of the pretrained GAN, andthen conduct iterative optimization to reconstruct images.Among them, mGANprior [13] attempts to optimize mul-tiple codes to improve the reconstruction quality. In or-der to reduce the gap between training and testing images,DGP [41] further jointly finetunes the generator and thelatent codes. However, these results struggle to faithfullyretain the local details, as the low-dimension latent codes

Page 3: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

FeatureExtractor

Pretrained GAN(Generative Color Prior)

𝑥𝑙

𝑥𝑖𝑛𝑣𝑟𝑔𝑏

𝑥𝑙𝑥𝑖𝑛𝑣𝑟𝑔𝑏

Warp

GAN Encoder

𝑥�𝑟𝑔𝑏

SPAD

E

Spat

ial

Corr

elat

ion

Colorization Network

z

SPAD

E

𝐹GAN

Figure 2: Overview of our framework. Given a grayscale image xl as input, our framework first produces the most relevantfeatures FGAN and an inversion image xrgbinv from a pretrained generative network as generative color priors. After that, wecalculate a spatial correlation matrix from xl and xrgbinv , and warp GAN features FGAN for alignment. The warped features areused to modulate the colorization network by spatially-adaptive denormalization (SPADE) layers. Finally, a vivid colorizedimage can be generated with the colorization network. In addition, controllable and diverse colorization results, together withsmooth transitions could be achieved by adjusting latent code z.

without spatial information are insufficient to guide the col-orization. In contrast, our method with spatial modulationenables prior incorporation on multi-resolution spatial fea-tures to achieve high texture faithfulness. In addition, ourmethod is feed-forward and does not require expensive iter-ative optimization for each instance. Recent works on facerestoration [51, 54] also utilize generative priors for restora-tion and colorization.

3. Method3.1. Overview

With a grayscale image xl ∈ R1×H×W as input, theobjective of colorization is to predict the two missingcolor channels a and b in the original color image xlab ∈R3×H×W , where l and a, b represent the luminance andchrominance in CIELAB color space, respectively. In thiswork, we aim to leverage rich and diverse color priors en-capsulated in a pretrained generative network to guide thecolorization. The overview of our framework is shown inFigure 2. It mainly consists of a pretrained GAN G, a GANEncoder E , and a Colorization Network C.

Given the grayscale image xl, GAN Encoder E is re-sponsible for mapping xl into a latent code z, which canserve as an input of G. The pretrained GAN G then gen-erates the most relevant color image (denoted as inversionimage xrgbinv) to the grayscale input. In addition to using theinversion image directly, the prior features FGAN from in-termediate layers are more informative and are leveraged

to guide the colorization. However, these ‘retrieved’ fea-tures usually do not spatially align with the grayscale input.To solve this, we first perform an alignment step to warpFGAN based on the correspondence between xl and xrgbinv .After that, we use the aligned GAN features to modulatethe multi-scale spatial features in Colorization Network Cwith effective SPADE (spatially-adaptive denormalization)layers [42], producing the final colorized image xrgb.

3.2. Generative Color Prior and GAN Encoder

Recently, GAN has achieved tremendous advances in un-supervised image generation. Once trained on an adequatenumber of images, GAN can map arbitrary random noisesinto realistic and high-resolution images with impressivedetails and rich colors, which motivates us to employ pre-trained GAN models to assist the colorization task. As il-lustrated above, we need to find a latent code z so that Gcan take it as the input and produce the GAN features FGANfrom its intermediate layers which are most relevant to xl.To achieve this goal, we constrain the generated image fromcode z to have similar content with xl but contains richchrominance information. The process of finding a certainz corresponding to a target image is termed as GAN inver-sion. A typical inversion way is to randomly initialize z andoptimize it to reduce the difference between the generatedimage and the target one. However, such a method requiresexpensive iterative optimization at inference and thus is notapplicable for real-world scenarios. Instead, we choose totrain a GAN encoder E to map arbitrary grayscale images

Page 4: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

into latent codes, i.e.,

z = E(xl),

xrgbinv, FGAN = G(z),(1)

where G(z) simultaneously outputs the inversion imagexrgbinv and the intermediate GAN features FGAN.

Discussion. It is well known that the optimization-basedinversion methods could generate more accurate inversionimages than encoder-based inversion methods [62, 64].However, in our colorization scenarios, we do not need suchprecise and high-quality inversion images. A coarse inver-sion result already contains sufficient color information andthe colorization network (illustrated in the next section) willfurther refine the corresponding inversion features. In addi-tion, we use the GAN features as extra information to guidethe colorization, which could provide complementary infor-mation to the inversion image.

3.3. Colorization Network

The colorization network C consists of two downsam-pling blocks, six residual blocks and two upsamplingblocks, which proves to be an effective architecture usedin many image-to-image translation tasks [50]. It takes thegrayscale image xl as input and predicts two missing colorchannels xab. A color image xlab can then be obtainedby concatenating the luminance channel xl and the chromi-nance channels xab. As the two color spaces RGB and LABcan be converted to each other, xrgb can also be obtained.

In order to use the prior color features FGAN to guidethe colorization, and to better preserve the color informa-tion of FGAN, we use spatially-adaptive denormalization(SPADE) [42] to modulate the colorization network. How-ever, FGAN can not be directly used to modulate network Cuntil they are spatially aligned with the input. Taking the de-picted image in Figure 2 as an example, the inversion imagexrgbinv contains all the semantic components that appeared inxl (i.e., the hen, soil and the weeds), but it is not spatiallyaligned with the input image.

Before we introduce the align operation, it is worth men-tioning that the prior color features FGAN are multi-scalefeatures, and the features from a certain scale are used tomodulate the layers of the corresponding scale in C. Forsimplicity, we only use a specific scale in FGAN to illustratethe align operation, denoted as F s

GAN ∈ RC×H′×W ′.

In the alignment module, we first use two feature ex-tractors with a shared backbone (denoted as FL→S andFRGB→S , respectively) to project xl and xrgbinv to a sharedfeature space S, obtaining the feature maps FL→S(xl)

and FRGB→S(xrgbinv). After that, we use a non-local op-eration [50] to calculate the correlation matrix M ∈

RH′W ′×H′W ′between the two feature maps. M(u, v) de-

notes the similarity between the input and the GAN featuresin position u and v. This operation is widely adopted to cal-culate dense semantic correspondence between two featuremaps [32, 56, 57]. Finally, we use the correlation matrixMto warp F s

GAN and obtain the aligned GAN features at scales, which are then used to modulate corresponding layers inC at scale s.

3.4. Objectives

GAN inversion losses. As we have the ground truth color-ful images xrgb during training, we can train GAN encoderE by minimizing the difference between xrgbinv and xrgb. In-stead of directly constraining the pixel values of xrgbinv andxrgb, we choose to minimize the discrepancy between theirfeatures extracted by the pre-trained discriminator Dg of G,as this will make the GAN inversion easier to optimize (es-pecially for BigGAN) [9, 41]. Formally, the discriminatorfeature loss Linv ftr is defined as:

Linv ftr =∑l

∥∥∥Dgl (xrgbinv)−Dg

l (xrgb)∥∥∥1, (2)

where Dgl represents the feature map extracted at l-th layer

from Dg .To keep z within a reasonable range, we also add a

L2 norm Linv reg , otherwise, the GAN inversion withgrayscale input will lead to unstable results:

Linv reg =1

2‖z‖2 . (3)

Adversarial loss. We adopt the adversarial loss to help thecolorization images looking more realistic. To stabilize theadversarial training, we employ the loss functions in LS-GAN [36]:

LDadv = E[(Dc(xlab)− 1)2] + E[(Dc(xlab))2],

LGadv = E[(Dc(xlab)− 1)2],(4)

where DC is the discriminator to discriminate the coloriza-tion images xlab generated from C and color images xlab

from real world. C and DC are trained alternatively withLGadv and LDadv , respectively.Perceptual loss. To make the colorization image perceptualplausible, we use the perceptual loss introduced in [25]:

Lperc =∥∥φl(xrgb)− φl(xrgb)∥∥2 , (5)

where φl represents the feature map extracted at l-th layerfrom a pretrained VGG19 network. Here ,we set l =relu5 2.Domain alignment loss. To ensure that the grayscale im-age and inversion image are mapped to a shared feature

Page 5: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

space in correlation calculation, we adopt a domain align-ment loss [57]:

Ldom =∥∥FL→S(xl)−FRGB→S(xrgb)

∥∥1. (6)

Contextual Loss. The contextual loss [37] can measurethe similarity between two non-aligned images. This lossterm was originally designed for image translation taskswith non-aligned data, and then extended to exemplar-basedcolorization [56] to encourage colors in the output imageto be close to those in the reference image. In this paper,we also adopt this loss to encourage the colorization imagexrgb to be relevant to the inversion image xrgbinv . Formally,the contextual loss Lctx is defined as:

Lctx =∑l

ωl

(− log(CX(φl(x

rgb), φl(xrgbinv)))

), (7)

where CX denotes the similarity metric between two featuresand we refer readers to [37] for more details. We use thelayer l = relu{3 2, 4 2, 5 2} from the pretrained VGG19network with weight ωl = {2, 4, 8} to calculate Lctx. Weset larger loss weights for deeper features to guide the net-work to focus more on the semantic similarity instead ofpixel-wise color closeness.Full objective. The full objective to train the GAN encoderE is formulated as

LE = λinv ftrLinv ftr + λinv regLinv reg. (8)

The full objective to train the colorization network C is for-mulated as

LC =λdomLdom + λpercLperc

+ λctxLctx + λadvLGadv.(9)

The full objective to train the discriminatorDC for coloriza-tion is formulated as

LDC= λadvLDadv. (10)

In the above loss objectives, λinv ftr, λinv reg, λdom,λperc,λctx and λadv are hyper-parameters that control the relativeimportance of each loss term.

4. Experiments4.1. Implementation Details

We conduct our experiments on ImageNet [45], a multi-class dataset, with a pretrained BigGAN [3] as the genera-tive color prior. We train our method with the official train-ing set. All the images are resized to 256× 256.

Our training process consists of two stages. In the firststage, we freeze the pretrained BigGAN generator and traina BigGAN encoder with LE . We utilize the last three lay-ers of BigGAN discriminator to calculate the feature loss,

where λinv ftr is set to 1.0 and λinv reg is set to 0.0125. Wetrain the model for totally 10 epochs with an initial learningrate of 1e−5.

In the second stage, the BigGAN encoder is frozen andthe rest of networks are trained end-to-end. We set the λpercto 1e−3, λadv to 1.0, λdom to 10.0 and λctx to 0.1. We trainthe model for totally 10 epochs with an initial learning rateof 1e−4. In both two stages, we set the batch size to 8per GPU and decay the learning rate linearly to 0. All themodels are trained with the Adam [29] optimizer.

The BigGAN encoder has a similar structure to the Big-GAN discriminator, while the class embedding is injectedinto the BigGAN encoder following the BigGAN generator.To reduce the learnable parameters ofFL→S andFRGB→S ,their backbones are shared. The discriminator for coloriza-tion Dc is a three-scale PatchGAN network [23, 50] withspectral normalization [40].

4.2. Comparisons with Previous Methods

To evaluate the performance of our method, we compareour results to other state-of-the-art automatic colorizationmethods including CIC [58], ChromaGAN [48], DeOldify1

and InstColorization [47]. We re-implement CIC and Chro-maGAN in PyTorch based on the original paper.Quantitative comparison. We test all the methods on theImageNet validation set of 50,000 images. The quantita-tive results on five metrics are reported in Table 1. FrechetInception Score (FID) [18] measures the distribution sim-ilarity between the colorization results and the ground truthcolor images. Our method achieves the lowest FID, indicat-ing that our method could generate colorization results withbetter image quality and fidelity. Colorfulness Score [16]reflects the vividness of generated images. Our method ob-tains a better colorfulness score (2nd column) and closestcolorfulness score to the ground-truth images (3rd column).The highest colorfulness score of CIC is an outlier, alsoobserved in [56]. This is probably because CIC encour-ages rare colors in the loss function, and the rare colors aremisjudged as vivid colors by this metric. We also providePSNR and SSIM for reference. However, it is well-knownthat such pixel-wise measurements may not well reflect theactual performance [4, 17, 39, 47, 48, 60], as plausible col-orization results probably diverge a lot from the originalcolor image.Qualitative Comparison. As shown Figure 3, compared toCIC, ChromaGAN and Deoldify, our results tend to obtainmore natural and vivid colors. For instance, in contrast toour vivid green, the broccoli color (Column 5) of the othermethods looks unnatural and yellowish. In addition, ourmethod generates better results with regards to consistenthues. In column 1, we can observe inconsistent chromatic-ities in the results of Deoldify and InstColorization, where

1https://github.com/jantic/DeOldify

Page 6: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

GT

Input

CIC

Chrom

aGAN

Ours

Deoldify

InstColorization

Figure 3: Visual comparisons with previous automatic colorization methods. Our method is able to generate more naturaland vivid colors.

Table 1: Quantitative comparison. ∆Colorful denotes theabsolute colorfulness score difference between the coloriza-tion images and the ground truth color images.

FID↓ Colorful↑ ∆Colorful↓ PSNR↑ SSIM↑

CIC 19.71 43.92 5.57 20.86 0.86ChromaGAN 5.16 27.49 10.86 23.12 0.87DeOldify 3.87 22.83 15.52 22.97 0.91InstColor 7.36 27.05 11.30 22.91 0.91Ours 3.62 35.13 3.22 21.81 0.88

the dress is yellow and blue at the same time. Though Chro-maGAN has consistent tones, it fails to capture the detailsby only mapping overall blue tones to the image, leadingto the blue artifacts on the lady’s face. Instead, our methodsuccessfully maintains the consistent tone and captures thedetails as shown in Column 1. Furthermore, our method canyield more diverse and lively colors for each small instanceas shown in column 7, while InstColorization fails to detectsuch small objects and inherently degrades to use its globalimage colorization network, whose results are dim and lessvivid.

Page 7: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Figure 4: Boxplots of user preferences for different meth-ods. Green dash lines represent the means. Our methodgot a significantly higher preference rate by users than othercolorization methods.

User study. In order to better evaluate the subjective qual-ity (i.e., vividness and diverseness of colors), we conducta user study to compare our method with the other state-of-art automatic colorization methods, i.e., CIC [58], Chro-maGAN [48], DeOldify and InstColorization [47]. We re-cruited 30 participants with normal or corrected-to-normalvision and without color blindness. We randomly select 70images from various categories of image contents (e.g., an-imals, plants, human beings, landscape, food, etc.). Foreach image, the grayscaled image is displayed at the left-most while five colorization results are displayed in a ran-dom manner to avoid potential bias. The participants arerequired to select the best-colorized image according to thecolorization quality in terms of vividness and diverseness.The boxplots are displayed in Figure 4. Our method ob-tained a significant preference ( 46.3%) by users than othercolorization methods (CIC 12.4%, ChromaGAN 11.7%,Deoldify 16.6%, InstColor 13.0%), demonstrating distinctadvantage on producing natural and vivid results.

4.3. Ablation Studies

Generative color prior. Generative color prior plays animportant role in providing exemplar features for vivid andvarious colors. When we remove the pretrained GANs,our method degrades to a common automatic colorizationmethod without guidance. As shown in Figure 5, the vari-ant experiment without GAN prior generates dim and bleakcolors, while our method could produce bright and joyfulimages, under the guidance of colorful exemplars in GANdistributions. We could also observe a large drop in FID andColorfulness score without generative color prior (Table 2).Feature guidance vs. image guidance. Our method adoptsmulti-resolution intermediate GAN features to guide thecolorization. The design could fully facilitate the color anddetail information from GAN features. Compared to imageguidance, our results are capable of recovering more vividcolors faithful to their underlying classes (e.g., the trees on

Input w/o GAN prior w/ GAN prior GT

Input Image guidance Feature guidance GT

Input w/o alignment w/ alignment GT

Figure 5: Qualitative comparisons of ablation studies onGAN prior, feature guidance and spatial alignment. GANprior introduces brighter and more various colors. Multi-resolution feature guidance could produce more coherentand rich colors than image guidance. Spatial alignment mit-igates the misalignment between ‘retrieved’ GAN featuresand input gray images.

the beach, the grass on the mountains), as multi-resolutionfeatures contain more semantic and color information (Fig-ure 5 and Table 2).Spatial alignment. The prior features from pretrainedGANs probably misalign with the input images. To addressthis problem, we employ spatial alignment to align the GANfeatures, leading to fewer artifacts and coherent results forcolorization. As shown in Figure 5, the tree and square col-ors of our results are not affected by other objects.

4.4. Controllable Diverse Colorization

In order to achieve diverse colorization, many attemptshave been made. For automatic colorization methods, diver-sity is usually achieved by stochastic sampling [10, 14, 44].However, the results are barely satisfactory and difficult to

Page 8: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Table 2: Quantitative comparisons for ablation studies.∆Colorful denotes the absolute colorfulness score differ-ence between the colorization images and the ground truthcolor images.

Variants FID↓ Colorful↑ ∆Colorful↓

Full Model 3.62 35.13 3.22w/o Generative Color Prior 8.40 31.21 7.14Image Guidance 4.01 26.12 12.23w/o Spatial Alignment 4.59 31.94 6.41

Inversions

Input

Diverse

Figure 6: Our method could adjust the latent codes to ob-tain various inversion results, thus easily achieving diversecolorization results for the parrot.

control (Detailed comparisons are provided in the supple-mentary material). For exemplar-based colorization meth-ods [17, 53], it is time-consuming and challenging to findlarge amounts of reference images with various styles. Dif-ferent from previous methods, our method could achievediverse and controllable colorization in a much easier andnovel manner.

On the one hand, we can adjust the latent codes with dis-turbance to obtain various exemplars and their correspond-ing features. As shown in Figure 6, various inversion resultsare attained by adjusting the latent codes, and correspond-ingly, diverse colorization results of parrots are obtained.On the other hand, our method inherits the merits of inter-pretable controls of GANs [20, 24, 46, 49] and could at-tain controllable and smooth transitions by walking throughGAN latent space. Specifically, we employ an unsupervisedmethod [49] to find color-relevant directions, such as light-ing, saturation, etc. As depicted in Figure 7, smooth transi-tions on background colors, object colors, saturation couldbe achieved.

4.5. Limitations

Though our method could produce appealing results inmost cases, it still has limitations. When the input imageis not in the GAN distribution or GAN inversion fails, ourmethod degrades to common automatic colorization meth-ods and may result in unnatural and incoherent colors. Asshown in Figure 8, the human on the beach and the cac-tus are missing from the GAN inversion results. Thus, ourmethod cannot find corresponding color guidance, result-ing in unnatural colorization for these objects. This couldbe mitigated by improving the GAN inversion and commonautomatic colorization approaches in future works.

Input background color

Input object color

Input saturation

Figure 7: With the interpretable controls of GANs, ourmethod could attain controllable and smooth transitions bywalking through GAN latent space.

Inputs Inversion Warped Results GT

Figure 8: Limitations of our model. The human in the beachand the cactus are missing from the GAN inversion, result-ing in unnatural colors.

5. ConclusionIn this work, we have developed a framework to produce

vivid and diverse colorization results by leveraging genera-tive color priors encapsulated in pretrained GANs. Specif-ically, we ‘retrieve’ multi-resolution GAN features condi-tioned on the input grayscale image with a GAN encoder,and incorporate such features into the colorization processwith feature modulations. One could easily realize diversecolorization by simply adjusting the latent codes or con-ditions in GANs. Moreover, our method could also attaincontrollable and smooth transitions by walking through theGAN latent space.

Page 9: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

References[1] Rameen Abdal, Yipeng Qin, and Peter Wonka. Im-

age2stylegan: How to embed images into the stylegan latentspace? In ICCV, 2019.

[2] Hyojin Bahng, Seungjoo Yoo, Wonwoong Cho, David Kee-tae Park, Ziming Wu, Xiaojuan Ma, and Jaegul Choo. Col-oring with words: Guiding image colorization through text-based palette generation. In ECCV, 2018.

[3] Andrew Brock, Jeff Donahue, and Karen Simonyan. Largescale gan training for high fidelity natural image synthesis.arXiv preprint arXiv:1809.11096, 2018.

[4] Yun Cao, Zhiming Zhou, Weinan Zhang, and Yong Yu.Unsupervised diverse colorization via generative adversarialnetworks. In Joint European conference on machine learn-ing and knowledge discovery in databases, 2017.

[5] Huiwen Chang, Ohad Fried, Yiming Liu, Stephen DiVerdi,and Adam Finkelstein. Palette-based photo recoloring. ACMTOG, 34(4):139–1, 2015.

[6] Guillaume Charpiat, Matthias Hofmann, and BernhardScholkopf. Automatic image colorization via multimodalpredictions. In ECCV, 2008.

[7] Zezhou Cheng, Qingxiong Yang, and Bin Sheng. Deep col-orization. In ICCV, 2015.

[8] Alex Yong-Sang Chia, Shaojie Zhuo, Raj Kumar Gupta, Yu-Wing Tai, Siu-Yeung Cho, Ping Tan, and Stephen Lin. Se-mantic colorization with internet images. In SIGGRAPH,2011.

[9] Giannis Daras, Augustus Odena, Han Zhang, and Alexan-dros G Dimakis. Your local gan: Designing two dimensionallocal attention mechanisms for generative models. In CVPR,2020.

[10] Aditya Deshpande, Jiajun Lu, Mao-Chuang Yeh, Min JinChong, and David Forsyth. Learning diverse image coloriza-tion. In CVPR, 2017.

[11] Aditya Deshpande, Jason Rock, and David Forsyth. Learn-ing large-scale automatic image colorization. In ICCV, 2015.

[12] Yuki Endo, Satoshi Iizuka, Yoshihiro Kanamori, and Jun Mi-tani. Deepprop: Extracting deep features from a single im-age for edit propagation. In Computer Graphics Forum, vol-ume 35, pages 189–201, 2016.

[13] Jinjin Gu, Yujun Shen, and Bolei Zhou. Image processingusing multi-code gan prior. In CVPR, 2020.

[14] Sergio Guadarrama, Ryan Dahl, David Bieber, JonathonShlens, Mohammad Norouzi, and Kevin Murphy. PixColor:Pixel recursive colorization. In BMVC, 2017.

[15] Raj Kumar Gupta, Alex Yong-Sang Chia, Deepu Rajan,Ee Sin Ng, and Huang Zhiyong. Image colorization usingsimilar images. In ACM MM, 2012.

[16] David Hasler and Sabine E Suesstrunk. Measuring color-fulness in natural images. In Human vision and electronicimaging VIII, volume 5007, pages 87–95. International So-ciety for Optics and Photonics, 2003.

[17] Mingming He, Dongdong Chen, Jing Liao, Pedro V. Sander,and Lu Yuan. Deep exemplar-based colorization. ACM TOG,37(4):1–16, 2018.

[18] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner,Bernhard Nessler, and Sepp Hochreiter. Gans trained by atwo time-scale update rule converge to a local nash equilib-rium. In NeurIPS, 2017.

[19] Yi-Chin Huang, Yi-Shin Tung, Jun-Cheng Chen, Sung-WenWang, and Ja-Ling Wu. An adaptive edge detection basedcolorization algorithm and its applications. In ACM MM,2005.

[20] Erik Harkonen, Aaron Hertzmann, Jaakko Lehtinen, andSylvain Paris. Ganspace: Discovering interpretable gan con-trols. In NeurIPS, 2020.

[21] Satoshi Iizuka and Edgar Simo-Serra. Deepremaster: tem-poral source-reference attention networks for comprehensivevideo enhancement. ACM TOG, 38(6):1–13, 2019.

[22] Satoshi Iizuka, Edgar Simo-Serra, and Hiroshi Ishikawa. Letthere be color! ACM TOG, 35(4):1–11, 2016.

[23] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A.Efros. Image-to-image translation with conditional adver-sarial networks. In CVPR, 2017.

[24] Ali Jahanian, Lucy Chai, and Phillip Isola. On the ”steerabil-ity” of generative adversarial networks. In ICLR, 2020.

[25] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptuallosses for real-time style transfer and super-resolution. InECCV, 2016.

[26] Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen.Progressive growing of gans for improved quality, stability,and variation. In ICLR, 2018.

[27] Tero Karras, Samuli Laine, and Timo Aila. A style-basedgenerator architecture for generative adversarial networks. InCVPR, 2018.

[28] Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten,Jaakko Lehtinen, and Timo Aila. Analyzing and improvingthe image quality of stylegan. In CVPR, 2020.

[29] Diederik P. Kingma and Jimmy Ba. Adam: A method forstochastic optimization. In ICLR, 2015.

[30] Manoj Kumar, Dirk Weissenborn, and Nal Kalchbrenner.Colorization transformer. In ICLR, 2021.

[31] Gustav Larsson, Michael Maire, and GregoryShakhnarovich. Learning representations for automaticcolorization. In ECCV, 2016.

[32] Junsoo Lee, Eungyeup Kim, Yunsung Lee, Dongjun Kim,Jaehyuk Chang, and Jaegul Choo. Reference-based sketchimage colorization using augmented-self reference anddense semantic correspondence. In CVPR, 2020.

[33] Anat Levin, Dani Lischinski, and Yair Weiss. Colorizationusing optimization. ACM TOG, 23(3):689–694, 2004.

[34] Xiaopei Liu, Liang Wan, Yingge Qu, Tien-Tsin Wong,Stephen Lin, Chi-Sing Leung, and Pheng-Ann Heng. In-trinsic colorization. In SIGGRAPH, 2008.

[35] Peng Lu, Jinbei Yu, Xujun Peng, Zhaoran Zhao, and XiaojieWang. Gray2colornet: Transfer more colors from referenceimage. In ACM MM, 2020.

[36] Xudong Mao, Qing Li, Haoran Xie, Raymond Y.K. Lau,Zhen Wang, and Stephen Paul Smolley. Least squares gen-erative adversarial networks. In ICCV, 2017.

[37] Roey Mechrez, Itamar Talmi, and Lihi Zelnik-Manor. Thecontextual loss for image transformation with non-aligneddata. In ECCV, 2018.

Page 10: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

[38] Sachit Menon, Alexandru Damian, Shijia Hu, Nikhil Ravi,and Cynthia Rudin. Pulse: Self-supervised photo upsam-pling via latent space exploration of generative models. InCVPR, 2020.

[39] Safa Messaoud, David Forsyth, and Alexander G. Schwing.Structural consistency and controllability for diverse col-orization. In ECCV. 2018.

[40] Takeru Miyato, Toshiki Kataoka, Masanori Koyama, andYuichi Yoshida. Spectral normalization for generative ad-versarial networks. In ICLR, 2018.

[41] Xingang Pan, Xiaohang Zhan, Bo Dai, Dahua Lin,Chen Change Loy, and Ping Luo. Exploiting deep genera-tive prior for versatile image restoration and manipulation.In ECCV, 2020.

[42] Taesung Park, Ming-Yu Liu, Ting-Chun Wang, and Jun-YanZhu. Semantic image synthesis with spatially-adaptive nor-malization. In CVPR, 2019.

[43] Yingge Qu, Tien-Tsin Wong, and Pheng-Ann Heng. Mangacolorization. ACM TOG, 25(3):1214–1220, 2006.

[44] Amelie Royer, Alexander Kolesnikov, and Christoph Lam-pert. Probabilistic image colorization. In BMVC, 2017.

[45] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, San-jeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy,Aditya Khosla, Michael Bernstein, Alexander C. Berg, andLi Fei-Fei. ImageNet large scale visual recognition chal-lenge. IJCV, 115(3):211–252, 2015.

[46] Yujun Shen and Bolei Zhou. Closed-form factorization oflatent semantics in gans. In CVPR, 2021.

[47] Jheng-Wei Su, Hung-Kuo Chu, and Jia-Bin Huang. Instance-aware image colorization. In CVPR, 2020.

[48] Patricia Vitoria, Lara Raad, and Coloma Ballester. Chro-magan: Adversarial picture colorization with semantic classdistribution. In WACV, 2020.

[49] Andrey Voynov and Artem Babenko. Unsupervised discov-ery of interpretable directions in the gan latent space. 2020.

[50] Ting-Chun Wang, Ming-Yu Liu, Jun-Yan Zhu, Andrew Tao,Jan Kautz, and Bryan Catanzaro. High-resolution image syn-thesis and semantic manipulation with conditional GANs. InCVPR, 2018.

[51] Xintao Wang, Yu Li, Honglun Zhang, and Ying Shan. To-wards real-world blind face restoration with generative facialprior. In CVPR, 2021.

[52] Kun Xu, Yong Li, Tao Ju, Shi-Min Hu, and Tian-Qiang Liu.Efficient affinity-based edit propagation using kd tree. ACMTOG, 28(5):1–6, 2009.

[53] Zhongyou Xu, Tingting Wang, Faming Fang, Yun Sheng,and Guixu Zhang. Stylization-based architecture for fastdeep exemplar colorization. In CVPR, 2020.

[54] Tao Yang, Peiran Ren, Xuansong Xie, and Lei Zhang. Ganprior embedded network for blind face restoration in thewild. In CVPR, 2021.

[55] L. Yatziv and G. Sapiro. Fast image and video colorizationusing chrominance blending. IEEE TIP, 15(5):1120–1129,2006.

[56] Bo Zhang, Mingming He, Jing Liao, Pedro V. Sander, LuYuan, Amine Bermak, and Dong Chen. Deep exemplar-based video colorization. In CVPR, 2019.

[57] Pan Zhang, Bo Zhang, Dong Chen, Lu Yuan, and Fang Wen.Cross-domain correspondence learning for exemplar-basedimage translation. In CVPR, 2020.

[58] Richard Zhang, Phillip Isola, and Alexei A. Efros. Colorfulimage colorization. In ECCV, 2016.

[59] Richard Zhang, Jun-Yan Zhu, Phillip Isola, Xinyang Geng,Angela S. Lin, Tianhe Yu, and Alexei A. Efros. Real-time user-guided image colorization with learned deep pri-ors. ACM TOG, 36(4):1–11, 2017.

[60] Jiaojiao Zhao, Jungong Han, Ling Shao, and Cees G. M.Snoek. Pixelated semantic colorization. IJCV, 128(4):818–834, 2019.

[61] Jiaojiao Zhao, Li Liu, Cees GM Snoek, Jungong Han, andLing Shao. Pixel-level semantics guided image colorization.In BMVC, 2018.

[62] Jiapeng Zhu, Yujun Shen, Deli Zhao, and Bolei Zhou. In-domain gan inversion for real image editing. In ECCV, 2020.

[63] Jiapeng Zhu, Yujun Shen, Deli Zhao, and Bolei Zhou. In-domain gan inversion for real image editing. In ECCV, 2020.

[64] Jun-Yan Zhu, Philipp Krahenbuhl, Eli Shechtman, andAlexei A Efros. Generative visual manipulation on the natu-ral image manifold. In ECCV, 2016.

Page 11: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Supplementary1. GAN Inversion Results

We show the GAN inversion results in Fig. 9. It is ob-served that the GAN inversion could generate colorful im-ages that share the similar semantic contents to the gray in-puts.

2. Ablation Study of BigGAN InversionDespite great success of GAN inversion in

StyleGAN[63], optimizing an encoder-based BigGANinversion model is still a challenging task [9, 41]. Simplyoptimizing the latent code z leads to large L2 norm of latentcode z and low-quality inverted images. We conjecture thatthis is because the self-attention architecture in BigGANbrings optimization challenges. To tackle this issue, wedirectly add an extra L2 norm penalty to the latent codeand empirically found that it could alleviate this issueand generate plausible inversion results. Besides, We alsoemploy the discriminator feature loss Linv ftr instead ofthe perceptual loss trained on classification task, since thefeature space of pretrained discriminator is more coherentwith that of inversion images. Recently important progresshas been achieved in StyleGAN inversion task, some usefultricks are introduced to generate high-quality inversionimages. However, not all of those useful tricks are suitablefor BigGAN, e.g., we found in practice that the domain-guided encoder introduced in [63] will not bring benefitsin BigGAN. Hence, the inversion of the encoder-basedBigGAN still remains a challenge.

In addition to the discriminator feature loss and L2 normscheme, we also make some attempts to improve BigGANinversion but their results are no better than the aforemen-tioned method. Those attempts include adopting more dis-criminator layers to calculate the feature loss and introduc-ing pixel loss. The comparisons of these attempts can befound in Fig. 10. It is observed that neither adding morediscriminator layers nor introducing the pixel loss will gen-erate better inversion images, but result in blurry imageswith fewer details.

3. Comparison with GAN InversionAs GAN inversions also incorporate pretrained GANs as

a prior to guide the colorization process, we further com-pare the results of our method against results of a GAN in-version method - DGP [41]. Selected representative casesare presented in Figure 11 for a qualitative comparison. Asshown in the figure, we can observe that color boundariesare not clearly separated in the results of DGP. The blurredcolor boundaries produced by DGP are inevitable consid-ering the information loss during GAN inversion process.The low-dimensional latent codes generated in DGP fail to

encode the spatial information in the image, thereby DGPis neither capable of retaining the local details nor preserv-ing the shape of objects. On the other hand, our method,which employs a SPADE modulation to incorporate the spa-tial features, has obtained better results in terms of texturefaithfulness. Apart from improving colorization quality,our method also reduces the computational cost by adopt-ing straightforward feed-forward inference to avoid iterativeoptimization training process used in DGP.

4. More Diverse Colorization ResultsIn this section, we provide comparisons of diverse col-

orization with stochastic sampling based method PIC [44].PIC adopts an autoregressive network PixelCNN to modelthe full joint distribution of pixel color values, by samplingfrom this distribution, diverse colorizations can be obtained.As shown in Fig. 12 and Fig. 13, PIC usually generates spa-tially incoherent and unrealistic colorization results. Onthe contrary, our method does not suffer from these issues.Besides, our method could attain smooth and controllablecolorizations.

5. User study resultsWe show some cases from our user study in Fig. 14 - 37.

Our method got a significant preference by users than othercolorization methods (CIC [58], ChromaGAN [48], DeOld-ify and InstColorization [47]), showing distinct advantageon producing natural and vivid results.

Page 12: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Input Inversion GT Input Inversion GT

Figure 9: Visual results of BigGAN inversion. It is observed that the GAN inversion could generate colorful images thatshare the similar semantic contents to the gray inputs. Left Column: input gray-scale images. Middle Column: the GANinversion results. Right Column: ground-truth colorful images.

Page 13: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Input GT

Figure 10: Ablation study of loss functions in BigGAN inversion. Ldis7 denotes using more discriminator layers (i.e., thelast 7 layers) to calculate the feature loss. Ldis3 denotes using the last 3 discriminator layers to calculate the feature loss.Ldis3 +Lpix denotes using the last 3 discriminator layers and pixel loss. It is observed that neither adding more discriminatorlayers nor introducing the pixel loss will generate better inversion images, but result in blurry images with fewer details.Therefore, we choose Ldis3 as our default setting.

Input DGP Ours Input DGP Ours

Figure 11: Comparisons with DGP.

Page 14: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

PICOurs

PICOurs

PICOurs

PICOurs

Figure 12: Comparisons of diverse colorization with PIC (First row: PIC; Second row: Ours). Zoom in to see the spatiallyincoherent and unrealistic colorizations from PIC. Our method could generate spatially coherent and more natural colorizationresults with smooth control.

Page 15: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

PICOurs

PICOurs

PICOurs

PICOurs

Figure 13: Comparisons of diverse colorization with PIC (First row: PIC; Second row: Ours). Zoom in to see the spatiallyincoherent and unrealistic colorizations from PIC. Our method could generate spatially coherent and more natural colorizationresults with smooth control.

Page 16: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Figure 14: User preferences for different methods.From column 2-6: 10%(CIC) : 3%(ChromaGAN) : 13%(DeOldify) : 3%(InstColor) : 70%(Ours).

Figure 15: User preferences for different methods.From column 2-6: 10%(CIC) : 7%(ChromaGAN) : 23%(DeOldify) : 10%(InstColor) : 50%(Ours).

Figure 16: User preferences for different methods.From column 2-6: 3%(CIC) : 3%(ChromaGAN) : 30%(DeOldify) : 3%(InstColor) : 60%(Ours).

Figure 17: User preferences for different methods.From column 2-6: 7%(CIC) : 10%(ChromaGAN) : 20%(DeOldify) : 10%(InstColor) : 53%(Ours).

Figure 18: User preferences for different methods.From column 2-6: 0%(CIC) : 7%(ChromaGAN) : 0%(DeOldify) : 23%(InstColor) : 70%(Ours).

Page 17: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Figure 19: User preferences for different methods.From column 2-6: 7%(CIC) : 13%(ChromaGAN) : 17%(DeOldify) : 23%(InstColor) : 40%(Ours).

Figure 20: User preferences for different methods.From column 2-6: 0%(CIC) : 13%(ChromaGAN) : 3%(DeOldify) : 33%(InstColor) : 50%(Ours).

Figure 21: User preferences for different methods.From column 2-6: 17%(CIC) : 10%(ChromaGAN) : 27%(DeOldify) : 0%(InstColor) : 47%(Ours).

Figure 22: User preferences for different methods.From column 2-6: 7%(CIC) : 7%(ChromaGAN) : 0%(DeOldify) : 27%(InstColor) : 60%(Ours).

Figure 23: User preferences for different methods.From column 2-6: 3%(CIC) : 7%(ChromaGAN) : 13%(DeOldify) : 3%(InstColor) : 73%(Ours).

Page 18: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Figure 24: User preferences for different methods.From column 2-6: 13%(CIC) : 7%(ChromaGAN) : 0%(DeOldify) : 0%(InstColor) : 80%(Ours).

Figure 25: User preferences for different methods.From column 2-6: 7%(CIC) : 13%(ChromaGAN) : 17%(DeOldify) : 10%(InstColor) : 53%(Ours).

Figure 26: User preferences for different methods.From column 2-6: 7%(CIC) : 0%(ChromaGAN) : 20%(DeOldify) : 0%(InstColor) : 73%(Ours).

Figure 27: User preferences for different methods.From column 2-6: 3%(CIC) : 0%(ChromaGAN) : 23%(DeOldify) : 13%(InstColor) : 60%(Ours).

Figure 28: User preferences for different methods.From column 2-6: 0%(CIC) : 13%(ChromaGAN) : 13%(DeOldify) : 0%(InstColor) : 73%(Ours).

Page 19: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Figure 29: User preferences for different methods.From column 2-6: 13%(CIC) : 0%(ChromaGAN) : 10%(DeOldify) : 13%(InstColor) : 63%(Ours).

Figure 30: User preferences for different methods.From column 2-6: 0%(CIC) : 30%(ChromaGAN) : 0%(DeOldify) : 0%(InstColor) : 70%(Ours).

Figure 31: User preferences for different methods.From column 2-6: 0%(CIC) : 10%(ChromaGAN) : 7%(DeOldify) : 0%(InstColor) : 47%(Ours).

Figure 32: User preferences for different methods.From column 2-6: 3%(CIC) : 43%(ChromaGAN) : 0%(DeOldify) : 7%(InstColor) : 83%(Ours).

Figure 33: User preferences for different methods.From column 2-6: 7%(CIC) : 0%(ChromaGAN) : 3%(DeOldify) : 10%(InstColor) : 80%(Ours).

Page 20: arXiv:2108.08826v1 [cs.CV] 19 Aug 2021

Figure 34: User preferences for different methods.From column 2-6: 10%(CIC) : 3%(ChromaGAN) : 0%(DeOldify) : 0%(InstColor) : 87%(Ours).

Figure 35: User preferences for different methods.From column 2-6: 10%(CIC) : 23%(ChromaGAN) : 3%(DeOldify) : 0%(InstColor) : 63%(Ours).

Figure 36: User preferences for different methods.From column 2-6: 17%(CIC) : 7%(ChromaGAN) : 3%(DeOldify) : 3%(InstColor) : 70%(Ours).

Figure 37: User preferences for different methods.From column 2-6: 17%(CIC) : 13%(ChromaGAN) : 13%(DeOldify) : 3%(InstColor) : 53%(Ours).