scene adaptation for semantic segmentation using ... · scene adaptation for semantic segmentation...

6
Scene Adaptation for Semantic Segmentation using Adversarial Learning D. Di Mauro 1 , A. Furnari 1 , G. Patan` e 2 , S. Battiato 1 , G. M. Farinella 1 1 Department of Mathematics and Computer Science – University of Catania – Viale A. Doria 6 - Catania {dimauro,furnari,battiato,gfarinella}@dmi.unict.it 2 Park Smart s.r.l. – Corso Italia, 189 – Catania {giuseppe.patane}@parksmart.it Abstract Semantic Segmentation algorithms based on the deep learning paradigm have reached outstanding perfor- mances. However, in order to achieve good results in a new domain, it is generally demanded to fine-tune a pre- trained deep architecture using new labeled data coming from the target application domain. The fine-tuning proce- dure is also required when the domain application settings change, e. g., when a camera is moved, or a new camera is installed. This implies the collection and pixel-wise la- beling of images to be used for training, which slows down the deployment of semantic segmentation systems in real in- dustrial scenarios and increases the industrial costs. Tak- ing into account the aforementioned issues, in this paper we propose an approach based on Adversarial Learning to per- form scene adaptation for semantic segmentation. We frame scene adaptation as the task of predicting semantic segmen- tation masks for images belonging to a Target Scene Con- text given labeled images coming from a Source Scene Con- text and unlabeled images coming from the Target Scene Context. Experiments highlight that the proposed method achieves promising performances both when the two scenes contain similar content (i.e., they are related to two differ- ent points of view of the same scene) and when the observed scenes contain unrelated content (i.e., they account to com- pletely different scenes). 1. Introduction Semantic image segmentation has a key role in many in- dustrial applications including video surveillance and traf- fic analysis [16, 2, 15]. Despite semantic segmentation ap- proaches based on deep learning have shown good perfor- mance on a variety of tasks (e.g., instance object segmen- tation [5], scene segmentation [1], etc.) they usually need Figure 1: Three cameras monitoring different parts of a city. S and S’ are related to the same part of the city, but different points of view, whereas S” is an observation of a different part of the city. to be adapted to the considered application domain through a fine-tuning process in order to achieve reasonable results. The fine-tuning process requires the collection and label- ing of a large amount of domain-dependent data. In ad- dition, whenever the system setup changes (e.g., cameras are moved or new cameras are installed), the fine-tuning procedure, including the collection and labeling of new data, needs to be repeated. Since collecting and labeling data requires a significant effort, the classic fine-tuning ap- proach prevents industrial applications to scale up. As noted in [18], domain adaptation techniques can be employed in order to limit the amount of labeled data to be used to adapt a pre-trained semantic segmentation algorithm. In this paper, we consider a scenario in which a fixed camera is installed to monitor a given area (e.g., a crossroad or a parking lot). When a new camera looking at a differ- ent scene or looking at the same scene but from a different point of view is added to the camera network, the seman- tic segmentation algorithms need to be adapted to the new view through a fine-tuning procedure. We investigate the 1

Upload: others

Post on 14-Jul-2020

19 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Scene Adaptation for Semantic Segmentation using ... · Scene Adaptation for Semantic Segmentation using Adversarial Learning D. Di Mauro 1, A. Furnari , G. Patan`e 2, S. Battiato

Scene Adaptation for Semantic Segmentation using Adversarial Learning

D. Di Mauro 1, A. Furnari 1, G. Patane 2, S. Battiato 1, G. M. Farinella 1

1 Department of Mathematics and Computer Science – University of Catania – Viale A. Doria 6 - Catania{dimauro,furnari,battiato,gfarinella}@dmi.unict.it

2 Park Smart s.r.l. – Corso Italia, 189 – Catania{giuseppe.patane}@parksmart.it

Abstract

Semantic Segmentation algorithms based on the deeplearning paradigm have reached outstanding perfor-mances. However, in order to achieve good results in anew domain, it is generally demanded to fine-tune a pre-trained deep architecture using new labeled data comingfrom the target application domain. The fine-tuning proce-dure is also required when the domain application settingschange, e. g., when a camera is moved, or a new camerais installed. This implies the collection and pixel-wise la-beling of images to be used for training, which slows downthe deployment of semantic segmentation systems in real in-dustrial scenarios and increases the industrial costs. Tak-ing into account the aforementioned issues, in this paper wepropose an approach based on Adversarial Learning to per-form scene adaptation for semantic segmentation. We framescene adaptation as the task of predicting semantic segmen-tation masks for images belonging to a Target Scene Con-text given labeled images coming from a Source Scene Con-text and unlabeled images coming from the Target SceneContext. Experiments highlight that the proposed methodachieves promising performances both when the two scenescontain similar content (i.e., they are related to two differ-ent points of view of the same scene) and when the observedscenes contain unrelated content (i.e., they account to com-pletely different scenes).

1. Introduction

Semantic image segmentation has a key role in many in-dustrial applications including video surveillance and traf-fic analysis [16, 2, 15]. Despite semantic segmentation ap-proaches based on deep learning have shown good perfor-mance on a variety of tasks (e.g., instance object segmen-tation [5], scene segmentation [1], etc.) they usually need

Figure 1: Three cameras monitoring different parts of a city.S and S’ are related to the same part of the city, but differentpoints of view, whereas S” is an observation of a differentpart of the city.

to be adapted to the considered application domain througha fine-tuning process in order to achieve reasonable results.The fine-tuning process requires the collection and label-ing of a large amount of domain-dependent data. In ad-dition, whenever the system setup changes (e.g., camerasare moved or new cameras are installed), the fine-tuningprocedure, including the collection and labeling of newdata, needs to be repeated. Since collecting and labelingdata requires a significant effort, the classic fine-tuning ap-proach prevents industrial applications to scale up. As notedin [18], domain adaptation techniques can be employed inorder to limit the amount of labeled data to be used to adapta pre-trained semantic segmentation algorithm.

In this paper, we consider a scenario in which a fixedcamera is installed to monitor a given area (e.g., a crossroador a parking lot). When a new camera looking at a differ-ent scene or looking at the same scene but from a differentpoint of view is added to the camera network, the seman-tic segmentation algorithms need to be adapted to the newview through a fine-tuning procedure. We investigate the

1

Page 2: Scene Adaptation for Semantic Segmentation using ... · Scene Adaptation for Semantic Segmentation using Adversarial Learning D. Di Mauro 1, A. Furnari , G. Patan`e 2, S. Battiato

use of Adversarial Learning [4] to limit the amount of la-beled data required for such adaptation. Specifically, werequire that only unlabeled data, effortlessly collected usingthe new camera, have to be used for the adaptation. Thetask is framed as a Domain Adaptation problem in whichlabeled images come from the first camera of the Source do-main, and unlabeled data come from the second camera ofthe Target domain. Figure 11 illustrates the proposed sceneadaptation problem.

To perform the study, we created a dataset of syntheticimages using the CARLA Urban Driving Simulator [3].The dataset consists of images collected from 3 differentscene contexts, each acquired from 2 different views. Ascene adaptation method based on Adversarial Learning ishence proposed. The method is designed to perform Se-mantic Segmentation of images of the Target domain aftera training phase which exploits labeled images from Sourcedomain and unlabeled images from Target domain. The ex-periments show that the proposed method outperforms thebaselines and achieves results close to (or in some case bet-ter than) those obtained by a classic fine-tuning approach.

2. Related Work

Generative Adversarial Networks: since their introduc-tion in 2014 [4], Generative Adversarial Networks havebeen employed in a series of applications, including superresolution [10], next video frame prediction [13], genera-tive visual manipulation [20] and image-translation [8]. Ourwork is related to the method proposed in Zhu et al. [21]where the problem of image-translation using unpaired do-mains has been addressed. Specifically, given images fromtwo domains X and Y , two translation functions are mod-eled: F : X → Y and G : Y → X . The mappings arelearned using unpaired samples by enforcing a cycle con-sistency loss.Semantic Segmentation: Long et al. [12] proposed to ex-ploit existing classification CNNs to produce dense imagesegmentation masks. Classification networks are modifiedby substituting fully connected layers with convolutions.The low resolution feature map produced by network is up-sampled using dilated deconvolutions to match the resolu-tion of the input image. Badrinarayanan et al. [1] presentedSegNet. The network is composed by a stack of encodersfollowed by a corresponding stack of decoders. The up-sampling is performed by the decoder using max-pool lo-cation indexes generated by the encoder. Zhao et al. [19]proposed to exploit global contextual information by con-text aggregation through a pyramid pooling module (PSP-Net). The considered global prior representation is effectiveto produce good quality results on the scene parsing task.

13D Model from http://s3-ap-southeast-2.amazonaws.com/launceston/atlas/index.html.

F

F G

G

Fake

RealD

Fake

RealD

Semantic Segmentation Loss

Reconstruction Loss

Reconstruction Loss

Adversarial Loss

Sour

ce D

omai

nTa

rget

Dom

ain

Figure 2: The proposed training architecture.

Domain Adaptation for Semantic Segmentation: previ-ous works investigated domain adaptation methods for se-mantic segmentation. Hoffman et al. [7] proposed to per-form global domain alignment using a novel semantic seg-mentation network with fully convolutional domain adver-sarial learning. Category specific adaptation is performedthrough a generalization of constrained weak learning, withexplicit transfer of the spatial layout from the source tothe target domains. In a successive work of Hoffman etal. [6] has been proposed to adapt representations both atthe pixel-level and feature-level through cycle-consistencywithout requiring aligned pairs. The model has been ap-plied to a variety of visual recognition and prediction set-tings, including the semantic segmentation of road scenes.Sankaranarayanan et al. [17] addresses domain adaptationfor semantic segmentation proposing a method consistingin an embedding network, a pixel-wise classifier, a genera-tor network and a discriminator network.All the aforementioned works mainly focused on imagescollected from moving cameras and on the adaptation fromsynthetic to real images. Differently from these methods,we investigate the task of scene adaptation, in which eachdomain accounts to a different scene acquired from a fixedcamera.

3. MethodWe propose an architecture which allows to train a se-

mantic segmentation network using labeled images fromthe Source domain and unlabeled images from the Targetdomain. At test time, the network can be used to predict se-mantic segmentation masks for images coming from bothdomains. The architecture is illustrated in Figure 2 andconsists of three components: the semantic network to betrained (F), a generative network (G), and a discriminatornetwork (D). The semantic network is responsible for pre-

Page 3: Scene Adaptation for Semantic Segmentation using ... · Scene Adaptation for Semantic Segmentation using Adversarial Learning D. Di Mauro 1, A. Furnari , G. Patan`e 2, S. Battiato

dicting semantic segmentation masks for both images be-longing to Source and Target domains. The ground truthsegmentation masks available for images belonging to theSource domain are used to provide direct supervision tonetwork F through a semantic segmentation loss. Since noground truth segmentation mask is available for images be-longing to the Target domain, supervision is in this case pro-vided by enforcing that the produced segmentation mask issufficient to reconstruct the input image. The enforcementis implemented through a reconstruction loss and an adver-sarial loss. This encourages the network to produce mean-ingful segmentation masks for both domains. The followingsections describe the proposed network and loss functionsin details.

3.1. Network Architecture

Semantic Segmentation Network: the semantic segmen-tation network F predicts segmentation masks for the in-put images. We implement this module using the architec-ture proposed Zhao et al. [19] (PSPNet) using a pre-trainedResNet backbone.Generative Network: the generative network G recon-structs the input images from the inferred segmentationmasks. The network takes over the pixelwise class scoresproduced by the Semantic Network F. To implement thegenerative network, we consider the architecture proposedby Johnson et al. [9] as implemented in [21].Discriminator Network: the discriminator network D isused to train the system using the paradigm of adversar-ial learning. This is done by training the discriminator todistinguish between real images and reconstructed ones.The semantic segmentation and generation networks arehence trained to “fool” the discriminator as described in [4].The discriminator network is implemented using a Patch-GANs [8, 11, 10] operating on 70 × 70 overlapping im-age patches. This patch-level discriminator architecture hasfewer parameters than a full-image discriminator and can beapplied to arbitrarily-sized images in a fully convolutionalfashion [8].

3.2. Loss Formulation

The model is trained by a composite loss which encour-ages the network to produce good semantic segmentationmasks as well as to reconstruct images correctly.Semantic Segmentation Loss: the semantic segmentationloss is defined as the cross entropy between the inferred pix-elwise probabilities and the ground truth per-pixel classes.The loss is computed only on images belonging to Sourcedomain, since no ground truth segmentation masks are as-sumed for Target domain. Let x be an image from Sourcedomain S, let F (x) be the set of scores produced by the se-mantic segmentation network and let S(F (x)) be softmaxof F (X), computed along class scores independently for

each pixel. Let y be the ground truth segmentation maskrelated to x indicating that pixel i belongs to class yi. Thesemantic segmentation loss is defined as:

LSemS(F ) = Ex∼pS(x)[−∑i

log(S(F (x))i,yi)] (1)

where pS denotes the distribution of data belonging to theSource domain, i iterates over the pixels of the training im-age x and S(F (x))i,yi denotes the probability predicted forclass yi at pixel i.Reconstruction Loss: following [21], we include a recon-struction loss to encourage the correct reconstruction of theinput images, starting from the class scores produced by thesemantic segmentation network. The loss is applied to bothimages belonging to the Source and Target domains. Weuse a L1 loss defined as follow:

LRec(G,F ) = Ex∼pST (x)[||G(F (x))− x||1] (2)

where pST denotes the distribution of data belonging toboth the Source and Target domains.Adversarial Loss: the adversarial loss is employed to en-courage the network to produce better reconstructions andovercome the common limitations of regression losses [14].Also this loss is applied to samples belonging to bothSource domain and Target domain. The loss is defined asfollows:

LGAN (G,F,D) = Ex∼pST (x)[log(D(x))]+

Ex∼pST (x)[log(1−D(G(F (x))))]. (3)

Overall Loss: the overall loss used to train the model isdefined as the sum of the three losses:

L(G,F,D) = LSemS(F ) + LRec(G,F )+

LGAN (G,F,D). (4)

4. DatasetsTo obtain labeled images of the same scene from differ-

ent points of views, we used CARLA Urban Driving Simu-lator [3] to create a synthetic dataset. Since the simulator istargeted to the collection of datasets for autonomous driv-ing, images can only be collected from moving cars. Imagesfrom fixed positions are hence collected placing the car ata predefined positions and setting up the cameras at differ-ent altitudes, pitch, yaw and roll angles relatively to the car.The car does not move in the simulation in order to obtain afixed point of view, while other objects (i.e. other vehiclesand pedestrians) are allowed to traverse the scene. Usingthe aforementioned procedure, we collected three episodesin three different scene contexts of the virtual city (referredto as “Scene 1”, “Scene 2” and “Scene 3”). Images from

Page 4: Scene Adaptation for Semantic Segmentation using ... · Scene Adaptation for Semantic Segmentation using Adversarial Learning D. Di Mauro 1, A. Furnari , G. Patan`e 2, S. Battiato

Scene 1 - View A (A1) Scene 1 - View B (B1)

Scene 2 - View A (A2) Scene 2 - View B (B2)

Scene 3 - View A (A3) Scene 3 - View B (B3)

Figure 3: Sample images and ground truth segmentation masks from the 6 data subsets.

each scene have been collected from two different points ofview (referred to as “View A” and “View B”) presentingscene overlap. This accounts to 6 image subsets denotedas “XY”, where X represent the view and Y represents thescene (e.g., A1, B2, etc.). Figure 3 reports some examplesfrom the 6 data subsets.We collected 5, 000 frames ad 1 fps for each episode-viewpair. The dataset contains 30, 000 frames in total. Eachimage has a resolution of 800 × 600 pixels. The time ofthe day and weather has been generated randomly for eachepisode. Each image of the dataset is associated to a groundtruth semantic segmentation mask produced by the simula-tor. The 13 annotated classes are: none, buildings, fences,other, pedestrians, poles, road-lines, roads, sidewalks, veg-etation, vehicles, walls and traffic signs.The collected dataset allows us to consider two differ-ent kinds of source-target domain pairs: 1) point ofview adaptation pairs, composed by two subsets from thesame scene context but with different views (e.g., A1 −B2, A2 − B2, etc.), 2) scene adaptation pairs, com-posed by two subsets belonging to different scene con-texts (e.g., A1 − A2, A2 − A3, etc.). The datasetis publicly available at http://iplab.dmi.unict.it/ParkSmartSceneAdaptation/avss.html.

5. Experimental Settings

We compare our method with respect to a baseline ob-tained training PSPNet on the Source domain and testing iton the Target domain, without adaptation. This baselinesis referred to as “No Adaptation” (NA) in the experiments.

The results are also compared with respect to a strong base-line obtained performing the fine-tune procedure, i.e., train-ing PSPNet directly on the Target domain using groundtruth annotations. This baselines is referred to as “Fine-Tuning” (FT) in the experiments. It should be noted that,making use of ground truth annotations, the FT baseline de-notes in fact the ideal performance, i.e. an upper bound, thatscene adaptation techniques should be able to achieve ide-ally. We consider two different experiments. In the former,we study adaptation with respect to different points of view.For this experiment, Source and Target domains are relatedto images acquired in the same scene from different pointsof view. In the latter, we study general scene adaptation. Inthis case, Source and Target domains are related to imagesacquired in different, non-overlapping scenes. All methodsare trained for 10 epochs. The weights of the best epochare used considered for the evaluation. We optimize thebaselines using Stochastic Gradient Descent with momen-tum 0.9, initial learning rate of 0.007 and (1 − i/3750)0.9

weight decay coefficient (where i is the current iteration,3750 total iteration) and batch size equal to 8. The proposedmethod is optimized using Adam, with initial learning rateequal to 0.0002 and batch size equal to 1. All the results areevaluated using standard measures for semantic segmenta-tion, as defined in [12]: Per Class Accuracy (cacc) and MeanIntersection over Union (miou).

6. Results

Table 1 reports the results of the first set of experimentsaimed to study point of view adaptation. The results are av-

Page 5: Scene Adaptation for Semantic Segmentation using ... · Scene Adaptation for Semantic Segmentation using Adversarial Learning D. Di Mauro 1, A. Furnari , G. Patan`e 2, S. Battiato

(a) Source (b) Target (c) NA (d) Proposed (e) FT (f) GT

Figure 4: Qualitative comparisons among the considered methods. The first an second rows show examples of point of viewadaptation, whereas the third row shows an example of scene context adaptation. (a) shows a sample image from the (labeled)source domain, (b) shows the processed image from the target domain, (c)-(e) show the results of the compared methods, and(f) shows the ground truth segmentation mask.

Table 1: Results of the point of view adaptation experiments. Best results between NA and Our are reported in bold. FTresults are reported in italic for reference and should be considered as an upper bound. * denotes the cases in which theproposed method outperforms the FT baseline.

Aver

age

Non

e

Bui

ldin

gs

Fenc

es

Oth

er

Pede

stri

ans

Pole

s

Roa

d-lin

es

Roa

ds

Side

wal

ks

Vege

tatio

n

Vehi

cles

Wal

ls

T.Si

gns

Per Class AccuracyNA 0.53 0.27 0.96 0.28 0.29 0.21 0.05 0.36 0.96 0.97 0.89 0.66 0.75 0.20

OUR 0.72* 0.90 0.95 0.55* 0.60* 0.53* 0.25* 0.48* 0.93 0.93 0.82 0.94* 0.82 0.68FT 0.69 0.96 0.98 0.50 0.46 0.21 0.22 0.44 0.98 0.98 0.85 0.84 0.88 0.73

Mean Intersection Over UnionNA 0.32 0.26 0.74 0.13 0.22 0.09 0.04 0.20 0.55 0.76 0.52 0.23 0.28 0.16

OUR 0.57 0.85 0.93 0.42 0.50* 0.13 0.20* 0.36* 0.86 0.88 0.71 0.26 0.71 0.63FT 0.64 0.91 0.98 0.42 0.42 0.19 0.17 0.34 0.95 0.92 0.74 0.78 0.83 0.70

Table 2: Results of the scene adaptation experiments. Best results between NA and Our are reported in bold. FT resultsare reported in italic for reference and should be considered as an upper bound. * denotes the cases in which the proposedmethod outperforms the FT baseline.

Aver

age

Non

e

Bui

ldin

gs

Fenc

es

Oth

er

Pede

stri

ans

Pole

s

Roa

d-lin

es

Roa

ds

Side

wal

ks

Vege

tatio

n

Vehi

cles

Wal

ls

T.Si

gns

Per Class AccuracyNA 0.27 0.53 0.09 0.07 0.02 0.05 0.03 0.13 0.84 0.67 0.41 0.61 0.00 0.04

OUR 0.34 0.69* 0.25 0.01 0.00 0.13 0.03 0.52* 0.77 0.50 0.49 0.81 0.00 0.18FT 0.68 0.23 0.99 0.52 0.70 0.38 0.14 0.43 0.99 0.98 0.94 0.89 0.94 0.65

Mean Intersection Over UnionNA 0.19 0.30 0.07 0.06 0.02 0.03 0.01 0.13 0.62 0.45 0.28 0.40 0.00 0.04

OUR 0.22 0.50* 0.14 0.01 0.00 0.02 0.02 0.40* 0.59 0.43 0.33 0.21 0.00 0.16*FT 0.47 0.22 0.90 0.21 0.40 0.15 0.12 0.33 0.88 0.88 0.54 0.71 0.71 0.07

eraged over a different experiments performed on the fol-lowing source-target image pairs: A1 − B1, A2 − B2,A3 − B3, B1 − A1, B2 − A2, B3 − A3. The proposedmethod outperforms the no adaptation (NA) baseline formost of the considered classes, while achieving similar re-

sults in the other cases. As can be noted, the proposed ap-proach also outperforms the fine-tuning (FT) baselines formany classes (see starred results in Table 1).

Table 2 summarizes the results related to the second setof experiments, aimed at assessing the scene adaptation ca-

Page 6: Scene Adaptation for Semantic Segmentation using ... · Scene Adaptation for Semantic Segmentation using Adversarial Learning D. Di Mauro 1, A. Furnari , G. Patan`e 2, S. Battiato

pabilities of the proposed method. The results are averagedover a different experiments performed on the followingsource-target image pairs: A1 − A2, A1 − A3, A2 − A1,A2 − A3, A3 − A1, A3 − A2. The results highlight thatgeneral scene adaptation is much more difficult than pointof view adaptation. Specifically, the proposed method tendsto outperform FT less often, while still allowing to obtainbetter results than the NA baseline.

Figure 4 shows some qualitative examples of the com-pared methods. It is worth to note that, the NA baselinetends to over-fit to the source domain, missing some detailssuch as the sky, the vegetation or the traffic lights, whilethe proposed method allows to recover most of these details– compare the proposed method with the Ground Truth la-bel (GT) in Figure 4. Moreover, in same cases, the pro-posed method allows to obtain results very similar to thoseachieved with the fine tuning procedure, while not using anylabel from the target domain (compare the proposed methodwith FT in the second row of Figure 4).

7. Conclusions

We have proposed a method to perform scene adaptationusing adversarial learning. The method improves the gener-alization of a semantic segmentation network by enforcingthe reconstruction of the input images from the generatedsemantic segmentation masks. Experiments show that theproposed method greatly reduces over-fitting in both pointof view and scene adaptation. Future work will explic-itly consider the geometric alignment of source and targetscenes to improve the results in the case of point of viewadaptation and, as a second step, we will try to apply suchresults on different domains.

References[1] V. Badrinarayanan, A. Handa, and R. Cipolla. Seg-

net: A deep convolutional encoder-decoder architecturefor robust semantic pixel-wise labelling. arXiv preprintarXiv:1505.07293, 2015.

[2] S. Battiato, G. M. Farinella, G. Gallo, and O. Giudice. On-board monitoring system for road traffic safety analysis.Computers in Industry, 98:208 – 217, 2018.

[3] A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, andV. Koltun. CARLA: An open urban driving simulator. InProceedings of the 1st Annual Conference on Robot Learn-ing, pages 1–16, 2017.

[4] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu,D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Gen-erative adversarial nets. In Advances in neural informationprocessing systems, pages 2672–2680, 2014.

[5] K. He, G. Gkioxari, P. Dollar, and R. Girshick. Mask r-cnn.In 2017 IEEE International Conference on Computer Vision(ICCV), pages 2980–2988, Oct 2017.

[6] J. Hoffman, E. Tzeng, T. Park, J.-Y. Zhu, P. Isola, K. Saenko,A. A. Efros, and T. Darrell. Cycada: Cycle-consistent adver-sarial domain adaptation. arXiv preprint arXiv:1711.03213,2017.

[7] J. Hoffman, D. Wang, F. Yu, and T. Darrell. Fcns in thewild: Pixel-level adversarial and constraint-based adapta-tion. arXiv preprint arXiv:1612.02649, 2016.

[8] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-imagetranslation with conditional adversarial networks. In Pro-ceedings of the IEEE Conference on Computer Vision andPattern Recognition, pages 1125–1134, 2017.

[9] J. Johnson, A. Alahi, and L. Fei-Fei. Perceptual losses forreal-time style transfer and super-resolution. In EuropeanConference on Computer Vision, pages 694–711. Springer,2016.

[10] C. Ledig, L. Theis, F. Huszar, J. Caballero, A. Cunningham,A. Acosta, A. Aitken, A. Tejani, J. Totz, Z. Wang, et al.Photo-realistic single image super-resolution using a genera-tive adversarial network. arXiv preprint, 2016.

[11] C. Li and M. Wand. Precomputed real-time texture syn-thesis with markovian generative adversarial networks. InEuropean Conference on Computer Vision, pages 702–716.Springer, 2016.

[12] J. Long, E. Shelhamer, and T. Darrell. Fully convolutionalnetworks for semantic segmentation. In Proceedings of theIEEE conference on computer vision and pattern recogni-tion, pages 3431–3440, 2015.

[13] W. Lotter, G. Kreiman, and D. Cox. Deep predictive cod-ing networks for video prediction and unsupervised learning.arXiv preprint arXiv:1605.08104, 2016.

[14] M. Mathieu, C. Couprie, and Y. LeCun. Deep multi-scalevideo prediction beyond mean square error. arXiv preprintarXiv:1511.05440, 2015.

[15] D. Ravı, M. Bober, G. M. Farinella, M. Guarnera, and S. Bat-tiato. Semantic segmentation of images exploiting dct basedfeatures and random forest. Pattern Recognition, 52:260–273, 2016.

[16] J.-F. Raymond. Traffic analysis: Protocols, attacks, designissues, and open problems. In Designing Privacy EnhancingTechnologies, pages 10–29, 2001.

[17] S. Sankaranarayanan, Y. Balaji, A. Jain, S. Lim, and R. Chel-lappa. Unsupervised domain adaptation for semantic seg-mentation with gans. CoRR, abs/1711.06969, 2017.

[18] E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell. Adversarialdiscriminative domain adaptation. CoRR, abs/1702.05464,2017.

[19] H. Zhao, J. Shi, X. Qi, X. Wang, and J. Jia. Pyramid sceneparsing network. In IEEE Conf. on Computer Vision andPattern Recognition (CVPR), pages 2881–2890, 2017.

[20] J.-Y. Zhu, P. Krahenbuhl, E. Shechtman, and A. A. Efros.Generative visual manipulation on the natural image mani-fold. In Proceedings of European Conference on ComputerVision (ECCV), 2016.

[21] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using cycle-consistent adversarial net-works. arXiv preprint arXiv:1703.10593, 2017.