arxiv:1612.09134v1 [cs.cv] 29 dec 2016refbase.cvc.uab.es/files/lxg2016.pdf · december 30, 2016...

22
Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016. From Virtual to Real World Visual Perception using Domain Adaptation – The DPM as Example Antonio M. L´opez Computer Vision Center (CVC) and Dpt. Ci` encies de la Computaci´o (DCC), Universitat Aut` onoma de Barcelona (UAB) [email protected] Jiaolong Xu CVC and DCC, UAB [email protected] Jos´ e L. G´ omez CVC and DCC, UAB [email protected] David V´ azquez CVC and DCC, UAB [email protected] Germ´ an Ros CVC and DCC, UAB [email protected] December 30, 2016 Abstract Supervised learning tends to produce more accu- rate classifiers than unsupervised learning in general. This implies that training data is preferred with an- notations. When addressing visual perception chal- lenges, such as localizing certain object classes within an image, the learning of the involved classifiers turns out to be a practical bottleneck. The reason is that, at least, we have to frame object examples with bounding boxes in thousands of images. A priori, the more complex the model is regarding its number of parameters, the more annotated examples are re- quired. This annotation task is performed by human oracles, which ends up in inaccuracies and errors in the annotations (aka ground truth) since the task is inherently very cumbersome and sometimes ambigu- ous. As an alternative we have pioneered the use of virtual worlds for collecting such annotations auto- matically and with high precision. However, since the models learned with virtual data must operate in the real world, we still need to perform domain adap- tation (DA). In this chapter we revisit the DA of a deformable part-based model (DPM) as an exemplify- ing case of virtual- to real-world DA. As a use case, we address the challenge of vehicle detection for driver assistance, using different publicly available virtual- world data. While doing so, we investigate questions such as: how does the domain gap behave due to virtual-vs-real data with respect to dominant object appearance per domain, as well as the role of photo- realism in the virtual world. 1 Need for Virtual Worlds Since the 90’s, machine learning has been an essential tool for solving computer vision tasks such as image classification, object detection, instance recognition, and (pixel-wise) semantic segmentation, among oth- ers [76, 13, 17, 54, 4]. In general terms, the best per- forming machine learning algorithms for these tasks are supervised; in other words, not only the raw data is required, but also annotated information, i.e. ground truth, must be provided to run the training arXiv:1612.09134v1 [cs.CV] 29 Dec 2016

Upload: others

Post on 29-Jul-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

From Virtual to Real World Visual Perception using Domain

Adaptation – The DPM as Example

Antonio M. LopezComputer Vision Center (CVC) and Dpt. Ciencies de la Computacio (DCC),

Universitat Autonoma de Barcelona (UAB)[email protected]

Jiaolong XuCVC and DCC, UAB

[email protected]

Jose L. GomezCVC and DCC, UAB

[email protected]

David VazquezCVC and DCC, UAB

[email protected]

German RosCVC and DCC, UAB

[email protected]

December 30, 2016

Abstract

Supervised learning tends to produce more accu-rate classifiers than unsupervised learning in general.This implies that training data is preferred with an-notations. When addressing visual perception chal-lenges, such as localizing certain object classes withinan image, the learning of the involved classifiers turnsout to be a practical bottleneck. The reason is that,at least, we have to frame object examples withbounding boxes in thousands of images. A priori,the more complex the model is regarding its numberof parameters, the more annotated examples are re-quired. This annotation task is performed by humanoracles, which ends up in inaccuracies and errors inthe annotations (aka ground truth) since the task isinherently very cumbersome and sometimes ambigu-ous. As an alternative we have pioneered the use ofvirtual worlds for collecting such annotations auto-matically and with high precision. However, sincethe models learned with virtual data must operate inthe real world, we still need to perform domain adap-

tation (DA). In this chapter we revisit the DA of adeformable part-based model (DPM) as an exemplify-ing case of virtual- to real-world DA. As a use case,we address the challenge of vehicle detection for driverassistance, using different publicly available virtual-world data. While doing so, we investigate questionssuch as: how does the domain gap behave due tovirtual-vs-real data with respect to dominant objectappearance per domain, as well as the role of photo-realism in the virtual world.

1 Need for Virtual Worlds

Since the 90’s, machine learning has been an essentialtool for solving computer vision tasks such as imageclassification, object detection, instance recognition,and (pixel-wise) semantic segmentation, among oth-ers [76, 13, 17, 54, 4]. In general terms, the best per-forming machine learning algorithms for these tasksare supervised; in other words, not only the rawdata is required, but also annotated information, i.e.ground truth, must be provided to run the training

arX

iv:1

612.

0913

4v1

[cs

.CV

] 2

9 D

ec 2

016

Page 2: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

protocol. Collecting the annotations has been basedon human oracles and collaborative software toolssuch as Amazon’s Mechanical Turk [48], LabelMe[53], etc. It is known, that human-based annotationis a cumbersome task, with ambiguities, and inaccu-racies. Moreover, not all kinds of ground truth canbe actually collected by relying on human annotators,e.g. pixel-wise optical flow and depth.

The non-expert reader can have a feeling of theannotation effort by looking at Fig. 1, where we cansee two typical annotation tasks, namely boundingbox (BB) based object annotations, and delineationof semantic contours between classes of interest. Inthe former case, the aim is to develop an object de-tector (e.g. a vehicle detector); in the latter, the aimis to develop a pixel-wise multi-class classifier, i.e. toperform the so-called semantic segmentation of theimage.

With the new century, different datasets were cre-ated with ground truth and put publicly available forresearch. Providing a comprehensive list of them isout of the scope of this chapter, but we can cite somemeaningful and pioneering examples related to twoparticular tasks in which we worked actively, namelypedestrian detection and semantic segmentation; bothin road scenarios for either advanced driver assistancesystems (ADAS) or autonomous driving (AD). Oneexample is the Daimler Pedestrian dataset [16], whichincludeds 3,915 BB-annotated pedestrians and 6,744pedestrian-free images (i.e. image-level annotations)for training, and 21,790 images with 56,492 BB-annotated pedestrians for testing. Another examplecorresponds to the pixel-wise class ground truth pro-vided in [7] for urban scenarios; giving rise to the well-known CamVid dataset which considers 32 semanticclasses (although only 11 are usually considered) andincludes 701 annotated images, 300 normally usedfor training and 401 for testing. A few years af-ter, the KITTI Vision Benchmark Suite [20] was anenormous contribution for the research focused onADAS/AD given the high variability of the providedsynchronized data (stereo images, LIDAR, GPS) andground truth (object bounding boxes, tracks, pixel-wise class, odometry).

In parallel to these annotation efforts and the cor-responding development of new algorithms (i.e. new

human-designed features, machine learning pipelines,image search schemes, etc.) for solving computer vi-sion tasks, deep leaning was finding its way to be-come the powerful tool that is today for solving suchtasks. Many researchers would point out [30] as amain breakthrough, since deep convolutional neu-ral networks (CNNs) showed an astonishing perfor-mance in the data used for the ImageNet Large-ScaleVisual Recognition Challenge (ILSVRC). ImageNet[14] contains over 15 million of human-labeled (usingMechanical Turk) high-resolution images of roughly22,000 categories. Thus, ImageNet was a gigantic hu-man annotation effort. ILSVRC uses a subset of Im-ageNet with about 1,000 images of 1,000 categories;overall, about 1.2M images for training, 50,000 forvalidation, and 150,000 for testing. Many deep CNNsdeveloped today rely on an ImageNet pre-traineddeep CNN which is modified or fine-tuned to solvea new task or operate in a new domain. The researchcommunity agrees in the fact that, in addition topowerful GPU hardware to train and test deep CNNs,having a large dataset with ground truth such as Im-ageNet is key for their success. In this line, more re-cently, it was released MS COCO dataset [34], whereper-instance object segmentation is provided for 91object types on 328,000 images, for a total of 2.5Mof labeled instances.

As a matter of fact, in the field of ADAS/AD wewould like to have datasets with at least the variety ofinformation sources of KITTI and the ground truthsize of ImageNet/COCO. However, when looking atthe ground truth of KITTI in quantitative terms,we can see that individually they are in the sameorder of magnitude than other ADAS/AD-orientedpublicly available datasets (e.g. see the number ofpedestrians BBs of KITTI and Daimler datasets, andthe number of pixel-wise annotated images of KITTIand CamVid). A proof of this need is the recentlyreleased Cityscapes dataset [12] which tries to gobeyond KITTI in several aspects. For instance, itincludes 5,000 pixel-wise annotated (stereo) imagescovering 30 classes and per-instance distinction, withGPS, odometry and ambient temperature as meta-data. In addition, it includes 20,000 more images butwhere the annotations are coarser regarding the de-lineation of the instance/class contours. This kind of

Page 3: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

Figure 1: Ground truth obtained by human annotation: left) framing the rectangular bounding box (BB)of vehicle instances; right) delineating the contours (silhouettes) between the different classes of interestcontained in the image, even at instance level.

dataset is difficult to collect since driving through 50cities covering several months and weather conditionswas required. Moreover, providing such a groundtruth can take from 30 to 90 minutes per image for ahuman oracle in case of fine-grained annotations anddepending on the image content.

For the semantic segmentation task, Cityscapesgoes one order of magnitude beyond KITTI andCamVid. However, it is far from the annotation num-bers of ImageNet and MS COCO. The main reasonis two-fold. On the one hand, data collection itself,i.e. Cityscapes images are collected from on-boardsystems designed for ADAS/AD not just downloadedfrom an internet source; moreover, metadata such asGPS and vehicle odometry is important, not to men-tion the possibility of obtaining depth from stereo.On the other hand, the annotations must be moreprecise since ultimately the main focus of ADAS/ADis on reducing traffic accidents. In any case, aswe mentioned before, other interesting ground truthtypes are not possible or really difficult to obtain byhuman annotation, e.g. pixel-wise optical flow anddepth (without active sensors); but eventually theseare important cues for ADAS/AD based on visualperception.

In this ADAS/AD context, and due to the difficul-ties and relevance of having large amounts of datawith ground truth for training, debugging and test-ing, roughly since 2008 we started to explore a differ-ent approach. In particular, the idea of using realisticvirtual worlds (e.g. based on videogames) for train-

ing vision-based perception modules. The advantageswere clear: (1) forcing the driving and data acquisi-tion situations needed; (2) obtaining different types ofpixel-wise ground truth (class ID, instance ID, depth,optical flow); (3) generating such data relatively fast(e.g. currently our SYNTHIA environment [50] cangenerate 10,000 images per hour with such groundtruths, using standard consumer hardware); etc. Ofcourse, such a proposal also came with doubts suchas can a visual model learned in virtual worlds oper-ate well in real-world environments?, and does thisdepend on the degree of photo-realism?. From our pi-oneering paper [36], where we used pedestrian detec-tion based on HOG/Linear-SVM as proof-of-concept,to our last work, i.e. SYNTHIA [50], where we haveaddressed pixel-wise semantic segmentation via deepCNNs, we have been continuously exploring the ideaof learning in virtual worlds to operate in real envi-ronments.

The use of synthetic data has attracted the atten-tion of other researchers too, and more recently spe-cially due to the massive adoption of deep CNNs toperform computer vision tasks and their data hun-gry nature. 3D CAD models have been used totrain visual models for pose estimation, object detec-tion and recognition, and indoor scene understand-ing [23, 1, 55, 59, 47, 57, 46, 56, 2, 10, 63, 44, 43,45, 26, 28, 52, 37, 62, 61, 41, 6, 27]; a virtual racingcircuit has been used for generating different typesof pixel-wise ground truth (depth, optical flow andclass ID) [25]; videogames have been used for train-

Page 4: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

ing deep CNNs with the purpose of semantic seg-mentation [49] and depth estimation from RGB [58];synthetic scenarios have been used also for evalu-ating the performance of different feature descrip-tors [29, 3, 74, 73, 75] and for training and test-ing optical and/or scene flow computation methods[39, 8, 42, 38], stereo algorithms [24], as well as track-ers [64], even using synthetic clones of real-world ar-eas of interest [18]; synthetic LIDAR-style data hasbeen used for object detection too [31, 32]; finally,virtual worlds are being used for learning high-levelartificial behavior such as playing Atari games [40],reproducing human behavior playing shooter games[35] and driving/navigating end-to-end [9, 82], evenlearning unwritten common sense [72, 83].

2 Need for Domain Adaptation

From the very beginning of our work, it was clear thatthere is a domain gap between virtual and real worlds.However, it was also clear that this was the casewhen using images coming from different (real) cam-era sensors and environments. In other words, thedomain gap is not a virtual-to-real issue, but rathera more general sensor-to-sensor or environment-to-environment problem [70, 71]. Other researchers con-firmed this fact too when addressing related but dif-ferent visual tasks than ours [66]. Since then, trainingvisual models in virtual worlds and applying domainadaptation techniques for their use in real-world sce-narios come hand-by-hand for us. In fact, more au-thors have followed the approach of performing someexplicit step of virtual- to real-world domain adapta-tion, without being an exhaustive list, the reader canaddress [32, 33, 63, 43] as illustrative examples.

We showed that virtual- to real-world domainadaptation is possible for holistic models based onthe HOG+LPB/Linear-SVM paradigm [68] as wellas on the Haar+EOH/AdaBoost one [69]. In theformer case, proof-of-concept experiments adaptingRGB-style synthetic images to far infrared ones (FIR)reported positive results too [60]. Moreover, for theDeformable Part-based Model (DPM) [17] we alsoproposed to use virtual worlds and domain adapta-tion [79, 77]. In most of the cases we focused on

supervised domain adaptation, i.e. a relatively fewamount of annotated target-domain data (i.e. fromthe real world in our case) was used to adapt themodel learned with source-domain data (from thevirtual world). For the holistic models we focused onmixing the source and target data collected via activelearning for model adaptation, we termed the corre-sponding feature space as cool world; while for DPMwe focused on using just the source-domain modeltogether with the target-domain data, i.e. withoutrevisiting the source-domain data. In terms of mod-ern deep CNNs, the former case would be similar tomixing source and target data in the mini-batcheswhile the latter case is more in the spirit of the so-called fine-tuning.

In the rest of this chapter we are going to focus onDPM because it was the state-of-the-art for objectdetection before the breakthrough of deep CNNs. Apriori it is a good proxy for deep CNNs regardingthe specific experiments we want to address, after alldeep CNNs eventually can require domain adaptationtoo [19, 67, 65, 11]. Obviously, being based on HOG-style features there is a point where much more datawould not really translate to better accuracy [81],so we will keep training data in the order of a fewthousands here. On the other hand, note that DPMcan be reformulated as a deep CNN [22] for end-to-end learning. Moreover, the domain adaptation tech-niques we proposed for DPM [77], can be used as coretechnology for hierarchical domain adaptation1 [78]as well as for weakly supervised incremental domainadaptation [80].

In particular, we are going to rely on our domainadaptation method for DPM termed as Structure-aware Adaptive Structural SVM (SA-SSVM), whichgave us the best performance in [77]. In this chapterwe compliment the experiments run in [77] mainly byaddressing questions such as the role of photo-realismin the virtual world, as well as how does the domaingap behave in virtual-vs-real data with respect to dom-inant object appearance per domain. Moreover, forthe sake of analyzing new use cases, instead of focus-ing on pedestrian detection using virtual data from

1With this technique we won the 1st pedestrian detectionchallenge of the KITTI benchmark suite, a part of the Recog-nition Meets Reconstruction Challenge held in ICCV’13.

Page 5: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

Figure 2: DPM for modeling pedestrians. There aretwo components (left, black background), and eachcomponent is encoded as a root and six parts (right,one component).

Half-Life 2 as in [77], here we focus on vehicle de-tection using different virtual-world datasets, namelyVirtual KITTI [18], SYNTHIA [50], and GTA [49].

3 Domain Adaptation for DPMin a Nutshell

DPM encodes the appearance of objects’ constituentparts together with a holistic object representationtermed as root. In contrast to other part mod-els, DPM allows the parts to be located at differ-ent positions with respect to the root. The plau-sible relative locations, known as deformations, arealso encoded. Both appearance and deformations arelearned. The appearance of the parts is learned atdouble the resolution than the root. The triplet root-parts-deformations is known as component. In orderto avoid too blurred models DPM allows to learn amixture of components. Different components use tocorrespond to very different object views or poses,specially when this implies very different aspect ra-tios of the corresponding root BB. See Fig. 2 for apictorial intuition.

In practice, a DPM is encoded as a vector w which

has to be learned. In the domain adaptation context,we term as wS the model learned with source-domaindata (e.g. with virtual-world data). Our SA-SSVMdomain adaptation method takes wS and relativelyfew annotated target-domain data (e.g. real-worlddata) to learn a new w model which is expected toperform better in the target domain. The reader isreferred to [77] for the mathematical technical de-tails of how SA-SSVM works. However, let us explainthe idea with the support of the example in Fig. 3;where wS consists of components: half body and fullbody, as well as persons seen from different view-points. Each component consists of root and parts(head, torso, etc). To adapt this DPM to a target

domain, we decompose it as wS = [wS1′, . . . ,wS

P′]′,

where P is the number of structures and u′ standsfor transpose of u. Note that each component, wS

p ,may contain both appearance and deformation pa-rameters (for roots only appearance). The decom-posed model parameters are adapted to the target do-main by different weights, denoted by βp, p ∈ {1, P};i.e. the SA-SSVM procedure allows domain adapta-tion for each of such structures separately by defin-ing ∆wp = wp − βpw

Sp , p ∈ {1, P}. In order to

learn these adaptation weights, we further introducea regularization term ‖β‖2 in the objective function,where β = [β1, . . . , βP ]′, and we use a scalar param-eter γ to control its relative penalty. Finally, C andthe ξi are just the standard terms of a SVM objec-tive function and N the number of target-domainsamples used for the adaptation. After optimizingfor the objective function (see mid box in Fig. 3),w = [w1

′, . . . ,wP′]′ is the domain adapted DPM.

4 Experimental Results

4.1 Datasets

As we mentioned before, we are going to focus on ve-hicle detection for ADAS/AD applications. We usethe training data of the KITTI car detection chal-lenge [21]; which is split into two sets, one for ac-tually training and the other for testing. Such atesting set will be the only one used here for thatpurpose, and we will follow the so-called moderate

Page 6: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

Figure 3: Domain Adaptation of DPM based on SA-SSVM (see main text for details).

Page 7: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

setting when considering which vehicles are manda-tory to detect. For training, we will consider fourmore datasets in addition to the mentioned split ofthe KITTI car detection challenge, thus, five in to-tal. Namely, the KITTI car tracking dataset [21], itssynthesized clone Virtual KITTI [18], SYNTHIA [50],and GTA [49]. Of course, SYNTHIA is a dataset withdifferent types of ground truth, so we selected a setof cars and images for our experiments. In the caseof GTA, we semi-automatically annotated with BBsa set of cars. Table 1 shows the number of samplesin each dataset. Figures 4, 5, 6, and 7, show imagessampled from KITTI-Det, KITTI Track with VirtualKITTI, SYNTHIA and GTA, respectively. VirtualKITTI and SYNTHIA are based on the same devel-opment framework, i.e. Unity3D2. We can see thatthe images from GTA are more photo-realistic thanthe ones used in SYNTHIA and Virtual KITTI. SYN-THIA images are not always corresponding to a for-ward facing on-board virtual camera as is the caseof Virtual KITTI and GTA. For more details aboutthe datasets the reader can refer to the correspondingpapers.

4.2 Protocol

In order to show the accuracy of the vehicle detec-tors we plot curves of false positive per image (FPPI)vs miss rate (MR) according to the Caltech protocol[15]; with an overlap of the 50% between detectionand ground truth BBs. For training and testing weonly consider moderate cases, which according to thedefinition given in the KITTI car detection challenge,are those vehicles non-occluded or just partially oc-cluded (maximum truncation: 30%), and with a BBheight ≥ 25 pixels. The vehicles mentioned in Tab.1 refer to moderate cases.

Regarding DPM we use three components, eachwith eight parts. Part locations are initialized as6 × 6 HOG-style cells (48 × 48 pixels) covering theroot (at its double resolution version). Note that,in contrast to [79, 77], here we have not exploredthe use of the pixel-wise vehicle masks (available forvirtual-world data) to provide a better initialization

2See unity3d.com

of part locations during DPM learning. Thus, real-and virtual-world training data are used equally forlearning source-domain DPMs.

Regarding the application of SA-SSVM we havefollowed the settings reported in [77] as producingthe best results. Namely, the adapted structurescorrespond to the root and parts, i.e. not to com-ponents; and we set γ = 0.08 and C = 0.001 (seeFig. 3). Since domain adaptation experiments (i.e.SA-SSVM based ones) require random sampling ofthe target domain training data, they are run threetimes and the mean FPPI-MR curve and standard-deviation based intervals are plotted (i.e. as in [77]but with three repetitions instead of five).

4.3 Experiments

According to the datasets listed in Tab. 1, we definethe set of source-domain datasets to be S = {KITTI-Track, Virtual KITTI, SYNTHIA, SYNTHIA-Sub,GTA}. The target-domain dataset is always KITTI-Det Test. KITTI-Det Train and KITTI-Det Testare coming from the same domain since they cor-respond to two splits we have done from the sameoriginal dataset. All the learned detectors are testedin KITTI-Det Test, and the difference among themis the data used for their training. Accordingly, thereported experiments are as follows:

• SRC: Training with a dataset s ∈ S.

• TAR-ALL: Training based on the full KITTI-Det Train.

• TARX: Training with a subset of random imagesfrom KITTI-Det Train, in particular, only usingthe 100X% of the images.

• SA-SSVM: Training with a dataset s ∈ S plusthe images used for the TARX shown in the sameplot.

Following this pattern, Fig. 8 shows results forX = 0.1 (i.e. 10%), Fig. 9 shows results for X = 0.5(i.e. 50%), and Fig. 10 shows results for X = 1 (i.e.ALL).

Page 8: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

Table 1: Used samples for each dataset. Images stands for the number of images and, from them, Vehiclesstands for the number of annotated vehicles using a bounding box. Negative samples are selected frombackground areas of the same images. KITTI-Det Test and KITTI-Det Train refer to two splits of thetraining set of the KITTI car detection training set. KITTI-Det Test is the testing set used in all theexperiments of this chapter, while the rest of datasets are used only for training. For KITTI Track andVirtual KITTI, we use sequences 1, 2, 6, 18, and 20 as the training datasets. SYNTHIA-sub refers to asubset randomly sampled from SYNTHIA.

KITTI-Det Test KITTI-Det Train KITTI-Track Virtual KITTI SYNTHIA SYNTHIA-Sub GTAImages 3163 3164 2020 1880 1313 675 580Vehicles 12894 12275 12950 6867 2052 1023 1054

Figure 4: Images sampled from KITTI-Det.

Page 9: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

Figure 5: Images sampled from KITTI-Track (left) and Virtual KITTI (right). Note how Virtual KITTI isa synthesized but realistic clone of KITTI-Track.

4.4 Discussion

The first observation comparing SRC and TAR-ALLresults (note that they are constant across figuressince do not depend on X) is that there is a largedomain gap. Since we would like to annotate as lessreal-world data as possible, let us start our analysisfor X = 0.1 (see Fig. 8).

The worst case is when SRC∈ {KITTI-Track, Vir-tual KITTI} since the average miss rate is ∼ 17points worse than for TAR-ALL. The best case iswhen SRC∈ {SYNTHIA, GTA}, where the gap is of∼ 12 points. Note that this is not related to thenumber of vehicle samples since GTA has ∼ 1/6 ofvehicles than Virtual KITTI for training, SYNTHIA∼ 1/3 than Virtual KITTI, and Virtual KITTI inturn contains ∼ 1/2 of the vehicles in KITTI-Trackand in KITTI-Det Test. An analogous behavior isseen when ranking the datasets by number of vehicle-free images. In any case, both ∼ 17 and ∼ 12 pointsare significant accuracy drops.

For going deeper in the analysis of what can bethe reason for the domain gap, we can comparethe results of KITTI-Track vs Virtual KITTI. We

see that they are basically equal. Since VirtualKITTI is a synthesized clone of KITTI-Track, wethink that the main reason of the accuracy drop isnot the virtual-to-real nature of the training images,but the typical vehicle poses and backgrounds re-flected in the training datasets, i.e. when comparingVirtual KITTI/KITTI-Track with KITTI-Det Train.In other words, KITTI-Det Train represents betterKITTI-Det Test since we built them from the samedata set. Note, in addition, that KITTI-Track comefrom the same camera sensor as KITTI-Det Trainand Test, which does not avoid the accuracy gap.Moreover, both SYNTHIA and GTA come from vir-tual worlds and still produce a detector that performsbetter than when using KITTI-Track.

We observe also that leaving out the ∼ 90% of theimages in KITTI-Det Train (X = 0.1) causes a dropin accuracy of ∼ 6 points. In other words, in thiscase once we have ∼ 316 manually annotated images(∼ 10% of KITTI-Det Train), annotating ∼ 2, 848more is required to push the DPM to its limit, whichis only ∼ 6 points better3. Active learning or ad hoc

3It is a fallacy to believe that, because good datasets are

Page 10: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

Figure 6: Images sampled from SYNTHIA dataset. Note that they are not always corresponding to a forwardfacing virtual camera on-board a car.

heuristics can be tried to alleviate such manual anno-tation effort. However, we observe that pre-trainingthe DPM with automatically collected virtual-worddata and using SA-SSVM for adapting the model,makes such ∼ 316 images already very valuable, sincein all cases the domain-adapted vehicle detector im-proves both the result of TAR0.1 and SRC (onlyvirtual-word data). We can see that the best caseis for SYNTHIA, which reduces the domain gap to∼ 2 points from ∼ 12 points, and improves the re-sult of TAR0.1 in ∼ 4 points. An additional obser-vation is that pre-training (SRC) the detectors withvirtual-world data also allows to use active learningtechniques as we did in [68] and/or ad hoc heuristicsas we did in [80] for annotating more complementaryimages (i.e. other more informative ∼ 316 ones) or

big, then big datasets are good [5].

collecting more but without human intervention (i.e.self-annotation). We have not done it here for thesake of simplicity, but it is reasonable to think thatthis would reduce the domain gap even more.

Figure 11 compares vehicle detections based onlyon the SYNTHIA samples we are using in this chap-ter, and the result of applying SA-SSVM to themwith TAR0.1, in both cases setting the threshold ofthe model classifier to operate in the FPPI=1 regime.Note how SA-SSVM allows to obtain better results.

TAR0.5 (X = 0.5; see Fig. 9) and TAR-ALL ba-sically show the same performance, so ∼ 1, 582 im-ages have been annotated without a reward in DPMperformance. Of course, although an annotation-training-test loop can be followed to avoid uselessvehicle annotations, a priori it is difficult to knowwhen to stop such manual annotations. On the otherhand, even using TAR0.5 data, starting with a pre-

Page 11: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

Figure 7: Images sampled from the GTA videogame.

Page 12: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 77.15%SA-SSVM = 68.17%TAR0.1 = 66.01%TAR-ALL = 59.92%

(a) KITTI-Track

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 77.19%TAR0.1 = 66.01%SA-SSVM = 64.26%TAR-ALL = 59.92%

(b) Virtual KITTI

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 71.86%TAR0.1 = 66.01%SA-SSVM = 62.12%TAR-ALL = 59.92%

(c) SYNTHIA

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 74.49%TAR0.1 = 66.01%SA-SSVM = 62.58%TAR-ALL = 59.92%

(d) SYNTHIA-Sub

10 -2 10 -1 10 0

false positives per image

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 71.07%TAR0.1 = 66.01%SA-SSVM = 64.39%TAR-ALL = 59.92%

(e) GTA

Figure 8: Results assuming X = 0.1 (see main text). In the box legend it is indicated the average miss ratefor each experiment. Thus, the lower the better.

Page 13: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 77.15%SA-SSVM = 62.12%TAR0.5 = 61.70%TAR-ALL = 59.92%

(a) KITTI-Track

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 77.19%TAR0.5 = 61.70%SA-SSVM = 60.88%TAR-ALL = 59.92%

(b) Virtual KITTI

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 71.86%TAR0.5 = 61.70%MIX = 60.71%TAR-ALL = 59.92%SA-SSVM = 57.89%

(c) SYNTHIA

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 74.49%TAR0.5 = 61.70%TAR-ALL = 59.92%SA-SSVM = 58.71%

(d) SYNTHIA-Sub

10 -2 10 -1 10 0

false positives per image

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 71.07%TAR0.5 = 61.70%TAR-ALL = 59.92%SA-SSVM = 58.30%

(e) GTA

Figure 9: Results assuming X = 0.5 (see main text). In the box legend it is indicated the average miss ratefor each experiment. Thus, the lower the better.

Page 14: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

10 -2 10 -1 10 0 10 1

false positives per image

0.2

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 77.15%SA-SSVM = 61.62%TAR-ALL = 59.92%

(a) KITTI-Track

10 -2 10 -1 10 0 10 1

false positives per image

0.2

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 77.19%TAR-ALL = 59.92%SA-SSVM = 60.18%

(b) Virtual KITTI

10 -2 10 -1 10 0 10 1

false positives per image

0.2

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 71.86%TAR-ALL = 59.92%MIX = 59.15%SA-SSVM = 56.20%

(c) SYNTHIA

10 -2 10 -1 10 0 10 1

false positives per image

0.2

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 74.49%TAR-ALL = 59.92%SA-SSVM = 57.14%

(d) SYNTHIA-Sub

10 -2 10 -1 10 0 10 1

false positives per image

0.2

0.3

0.4

0.5

0.64

0.8

1

mis

s ra

te

SRC = 71.07%TAR-ALL = 59.92%SA-SSVM = 56.97%

(e) GTA

Figure 10: Results assuming X = 1 (ALL; see main text). In the box legend it is indicated the average missrate for each experiment. Thus, the lower the better.

Page 15: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

Figure 11: Vehicle detections when operating in the FPPI=1 regime. Left: DPM based on the SYNTHIAdata considered in this chapter (SRC). Middle: Using the TAR0.1 version of KITTI-Det Train. Right:Adapting SRC by using TAR0.1 when applying SA-SSVM.

Page 16: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

trained model (SRC) on either SYNTHIA or GTAallows to improve the performance of TAR-ALL, inthe case of SYNTHIA by ∼ 2 points with respect toTAR0.5. Looking at SYNTHIA-Sub and GTA, whichhas a similar number of samples (see Tab. 1), we canargue that GTA could probably reach the same per-formance than SYNTHIA if we would have the doubleof GTA vehicles. In any case, what it is remarkable isthat it is more effective to have DPMs pre-trained invirtual worlds than just doubling the number of man-ually annotated target-domain images, i.e. at leastassuming a manual annotation procedure free of priorknowledge about the current vehicle detector.

Even for X = 1 (see Fig. 10), i.e. combining thedata used to train TAR-ALL and SRC, pre-trainingin virtual worlds is able to improve the performanceof TAR-ALL alone. SYNTHIA provides us ∼ 3points of improvement with respect to TAR-ALL, be-ing the overall best result. Using GTA as SRC even-tually can provide such improvement too (again, byextrapolation of its performance when comparing toSYNTHIA-Sub).

In summary, the results presented and discussed sofar reinforce the take home messages we highlightedin our previous works [36, 68, 77]; namely, combiningmodels/data pre-trained in virtual worlds with a rea-sonable but low amount of real-world data throughdomain adaptation, is a really practical paradigmworth to explore for learning different kinds of mod-els. As a matter of fact, according to the literaturereviewed in sections 1 and 2, nowadays this approachis being widely adopted by the computer vision com-munity.

Another interesting question that we did not ad-dressed before refers to the degree of photo-realism,i.e. if a higher degree would imply to learn more accu-rate models and eventually not even requiring domainadaptation. This is a very important question since aextreme photo-realism may require hours for render-ing a few images, while the degree of photo-realism ofthe virtual worlds presented here is achieved in realtime using a standard modern gamer PC.

In our previous works we already saw that domainadaptation was required even when you train and testwith real-world cameras. In other words, domain gapwas due to sensor differences (no matter if one of the

sensors operates in real or virtual worlds) and thenature of the scenario where train and test imagesare acquired (typical illumination, background, andpose/view of dynamic objects). Because of this, ourbelieve was that a more photo-realistic world wouldbe just another sensor, still different from real-world,and therefore domain gaps would persists. Note thatthe experiments presented in this chapter reinforcethis hypothesis: (1) using Virtual KITTI and KITTI-Track gives rise to SRC and domain-adapted detec-tors of similar performance in all the cases, i.e. de-spite the fact that KITTI-Track relies on the samereal-world sensor than KITTI-Det Train and Test,while Virtual KITTI consists of synthesized data; (2)moreover, despite the fact that GTA contains imagesmore photo-realistic than SYNTHIA, when using asimilar number of samples (SYNTHIA-Sub) we seethat the performance of the corresponding SRC andthe domain-adapted detectors is basically the same.

Recent works [41, 75] reinforce the idea that, oncea basic photo-realism is achieved (i.e. beyond Lam-bertian illumination and simplistic object materials),adding more and more photo-realism do not have arelevant impact. Thus, in our opinion from the evi-dences collected so far, Virtual KITTI and SYNTHIAare sufficiently photo-realistic for the tasks we are ad-dressing (i.e. vision-based object detection and im-age semantic segmentation).

Another interesting point of analysis is if it is bet-ter to just mixing virtual- and real-world data orfine-tuning a pre-trained model on virtual-world datawith real-world samples. The former is what wecalled cool world [70, 68], while SA-SSVM is an exam-ple of the later. Because of that we have run a MIXexperiment with SYNTHIA and TAR-ALL, whichcan be seen in Fig. 10(c). In this case, we have justmixed the data and run an standard DPM learningprocedure. Note that the result is ∼ 3 points worsethan using SA-SSVM. Moreover, the training time ofMIX is much longer than the one of SA-SSVM, sinceit uses samples from both domains and training fromscratch also requires more iterations to converge. Ifwe extrapolate these experiments to the deep CNNsparadigm, a priori we would think than fine-tuningis the proper approach. However, when working in[50, 51], i.e. in semantic segmentation based on deep

Page 17: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

CNNs, using the appropriate mini-batch scheme toweight the relevance of the samples as a function totheir domain (virtual or real), we obtained better per-formance than with fine-tuning. Therefore, regard-ing this topic, we have no clear conclusions yet. Ofcourse, the advantage of fine-tuning would be avoid-ing to revisit the source data; thus, this is a point tokeep researching.

Overall, our research and the research presentedso far by the computer vision community, led us toinsist in the adoption of the paradigm where virtualworlds and domain adaptation techniques are usedto train the desired models. Moreover, we think thatthe degree of photo-realism like the presented alreadyin datasets such as Virtual KITTI and SYNTHIA issufficient for this task. In addition, although in thischapter we have focused on DPM-based vehicle detec-tion, we think the conclusions can be extrapolated toother computer vision tasks where the visual appear-ance is important (e.g. object detection in generaland semantic segmentation). Of course, it is worthto note that at this moment the best results on theKITTI car detection challenge are dominated by ap-proaches based on deep CNNs, providing astonishinghigh performances in the moderate setting, far be-yond DPM approaches. Such benchmark seems tobe challenging enough for DPM, but still is a smallproportion of the real-world and this will be the realchallenge for deep CNNs. Therefore, we also thinkthat our conclusions will be extrapolated from DPMto other powerful models such as deep CNNs whenaddressing more challenging scenarios; note that inSect. 2 we have mentioned already that even deepCNNs require domain adaptation. On the otherhand, what is expected is that deep CNNs would re-quire less domain adaptation than DPM since theyare models with more capacity to generalize acrossdomains.

5 Conclusion

In this chapter we have shown how virtual worldsare effective for training visual models when com-bined with domain adaptation techniques. Althoughwe have focused on DPM and vehicle detection as

proof-of-concept, we believe that the conclusions ex-trapolate to other visual tasks based on more com-plex models such as deep CNNs. We have presentedresults which suggest that extreme photo-realism isnot necessary, i.e. the degree of photo-realism al-ready achieved in datasets such as Virtual KITTI andSYNTHIA is sufficient, provided domain adaptationwould be necessary even when relying on more photo-realistic datasets (here GTA).

Looking into the future, we think a best prac-tice would be to design sets of relatively controlledvirtual-world scenarios, designed to train, debug andtest visual perception and other AI capabilities (Vir-tual KITTI and SYNTHIA are examples of this). Inother words, with the knowledge accumulated so far,we do not bet for building a gigantic virtual world totry to avoid domain gap issues. This would be reallydifficult to build and handle. We prefer to pursue do-main adaptation to save any existing virtual-to-realworld gap. However, we think the research must gointo the direction of unsupervised domain adaptationfor allowing the systems trained in virtual worlds toself-adapt to real-world scenarios. An example in thisline is the approach we presented in [80], where man-ual annotations were not required to train a domainadapted pedestrian detector for an on-board movingcamera setting. However, this approach performs theadaptation off-line, which can be perfectly right formany applications (e.g. adapting pre-trained surveil-lance systems to different places), but the real chal-lenge is to do it on-line.

Acknowledgments Authors want to thank the next

funding bodies: the Spanish MEC Project TRA2014-

57088-C2-1-R, the People Programme (Marie Curie Ac-

tions) FP7/2007-2013 REA grant agreement no. 600388,

and by the Agency of Competitiveness for Companies of

the Government of Catalonia, ACCIO, the Generalitat de

Catalunya Project 2014-SGR-1506 and the NVIDIA Cor-

poration for the generous support in the form of different

GPU hardware units.

References

[1] Ankur Agarwal and Bill Triggs. A local basisrepresentation for estimating human pose from

Page 18: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

cluttered images. In Asian Conference on Com-puter Vision (ACCV), 2006.

[2] Mathieu Aubry, Daniel Maturana, Alexei Efros,Bryan Russell, and Josef Sivic. Seeing 3D chairs:exemplar part-based 2d-3d alignment using alarge dataset of CAD models. In IEEE Con-ference on Computer Vision and Pattern Recog-nition (CVPR), 2014.

[3] Mathieu Aubry and Bryan C. Russell. Under-standing deep features with computer-generatedimagery. In IEEE International Conference onComputer Vision (ICCV), 2015.

[4] Yoshua Bengio, Aaron Courville, and PascalVincent. Representation learning: A reviewand new perspectives. Transactions of Pat-tern Recognition and Machine Analyses (PAMI),35(8):1798–1828, 2013.

[5] Tamara L. Berg, Alexander Sorokin, GangWang, David A. Forsyth, Derek Hoiem, Ian En-dres, and Ali Farhadi. It’s all about the data.Proceedings of the IEEE, 98(8):1434–1452, 2010.

[6] Erik Bochinski, Volker Eiselein, and ThomasSikora. Training a convolutional neural networkfor multi-class object detection using solely vir-tual world data. In Advanced Video and Signal-based Surveillance (AVSS), 2016.

[7] Gabriel J. Brostow, Julien Fauqueur, andRoberto Cipolla. Semantic object classes invideo: A high-definition ground truth database.Pattern Recognition Letters, 30(20):88–89, 2009.

[8] Daniel J. Butler, Jonas Wulff, Garrett B. Stan-ley, and Michael J. Black. A naturalisticopen source movie for optical flow evaluation.In European Conference on Computer Vision(ECCV), 2012.

[9] Chenyi Chen, Ari Seff, Alain L. Kornhauser, andJianxiong Xiao. DeepDriving: Learning affor-dance for direct perception in autonomous driv-ing. In IEEE International Conference on Com-puter Vision (ICCV), 2015.

[10] Liang-Chieh Chen, Sanja Fidler, and RaquelYuille, Alan L. Urtasun. Beat the MTurkers:Automatic image labeling from weak 3D super-vision. In IEEE Conference on Computer Visionand Pattern Recognition (CVPR), 2014.

[11] Brian Chu, Vashisht Madhavan, Oscar Bei-jbom, Judy Hoffman, and Trevor Darrell. Bestpractices for fine-tuning visual classifiers tonew domains. In European Conference onComputer Vision (ECCV), Transferring andAdapting Source Knowledge in Computer Vision(TASK-CV), 2016.

[12] Marius Cordts, Mohamed Omran, SebastianRamos, Timo Rehfeld, Markus Enzweiler, Ro-drigo Benenson, Uwe Franke, Stefan Roth, andBernt Schiele. The Cityscapes dataset for seman-tic urban scene understanding. In IEEE Confer-ence on Computer Vision and Pattern Recogni-tion (CVPR), 2016.

[13] Navneet Dalal and Bill Triggs. Histograms of ori-ented gradients for human detection. In IEEEConference on Computer Vision and PatternRecognition (CVPR), pages 886–893, 2005.

[14] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li,Kai Li, and Li Fei-Fei. Imagenet: A large-scalehierarchical image database. In IEEE Confer-ence on Computer Vision and Pattern Recogni-tion (CVPR), 2009.

[15] Piotr Dollar, Christian Wojek, Bernt Schiele,and Pietro Perona. Pedestrian detection: anevaluation of the state of the art. Transactionsof Pattern Recognition and Machine Analyses(PAMI), 34(4):743–761, 2012.

[16] Markus Enzweiler and Dariu M. Gavrila.Monocular pedestrian detection: Survey and ex-periments. Transactions of Pattern Recognitionand Machine Analyses (PAMI), 31(12):2179–2195, 2009.

[17] Pedro F Felzenszwalb, Ross B Girshick, DavidMcAllester, and Deva Ramanan. Object de-tection with discriminatively trained part-based

Page 19: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

models. Transactions of Pattern Recognitionand Machine Analyses (PAMI), 32(9):1627–1645, 2010.

[18] Adrien Gaidon, Qiao Wang, Yohann Cabon, andEleonora Vig. Virtual worlds as proxy for multi-object tracking analysis. In IEEE Conferenceon Computer Vision and Pattern Recognition(CVPR), 2016.

[19] Yaroslav Ganin and Victor Lempitsky. Unsuper-vised domain adaptation by backpropagation. InInternational Conference on Machine Learning(ICML), pages 1180–1189, 2015.

[20] Andreas Geiger, Philip Lenz, Christoph Stiller,and Raquel Urtasun. Vision meets robotics:The KITTI dataset. International Journalof Robotics Research (IJRR), 32(11):1231–1237,2016.

[21] Andreas Geiger, Philip Lenz, and Raquel Urta-sun. Are we ready for autonomous driving? theKITTI vision benchmark suite. In IEEE Con-ference on Computer Vision and Pattern Recog-nition (CVPR), 2012.

[22] Ross Girshick, Forrest Iandola, Trevor Darrell,and Jitendra Malik. Deformable part models areconvolutional neural networks. In IEEE Confer-ence on Computer Vision and Pattern Recogni-tion (CVPR), pages 437–446, 2015.

[23] Kristen Grauman, Gregory Shakhnarovich, andTrevor Darrell. Inferring 3D structure with astatistical image-based shape model. In IEEEInternational Conference on Computer Vision(ICCV), 2003.

[24] Ralf Haeusler and Daniel Kondermann. Synthe-sizing real world stereo challenges. In GermanConference on Pattern Recognition (GCPR),2013.

[25] Haltakov Haltakov, Christian Unger, and Slobo-dan Ilic. Framework for generation of syntheticground truth data for driver assistance applica-tions. In German Conference on Pattern Recog-nition (GCPR), 2013.

[26] Ankur Handa, Viorica Patraucean, VijayBadrinarayanan, Simon Stent, and RobertoCipolla. SynthCam3D: Semantic understand-ing with synthetic indoor scenes. CoRR,arXiv:1505.00171, 2015.

[27] Ankur Handa, Viorica Patraucean, Vijay Badri-narayanan, Simon Stent, and Roberto Cipolla.Understanding real world indoor scenes withsynthetic data. In IEEE Conference on Com-puter Vision and Pattern Recognition (CVPR),2016.

[28] Hironori Hattori, Vishnu Naresh Boddeti,Kris M. Kitani, and Takeo Kanade. Learningscene-specific pedestrian detectors without realdata. In IEEE Conference on Computer Visionand Pattern Recognition (CVPR), 2015.

[29] Biliana Kaneva, Antonio Torralba, and WilliamT. Freeman. Evaluation of image features usinga photorealistic virtual world. In IEEE Interna-tional Conference on Computer Vision (ICCV),2011.

[30] Alex Krizhevsky, Ilya Sutskever, and Geoff Hin-ton. ImageNet classification with deep Convo-lutional Neural Networks. In Annual Confer-ence on Neural Information Processing Systems(NIPS), 2012.

[31] Kevin Lai and Dieter Fox. 3D laser scan classi-fication using web data and domain adaptation.In Robotics: Science and Systems, 2009.

[32] Kevin Lai and Dieter Fox. Object recognitionin 3D point clouds using web data and domainadaptation. International Journal of RoboticsResearch (IJRR), 29(8):1019–1037, 2010.

[33] Wenbin Li and Mario Fritz. Recognizing mate-rials from virtual examples. In European Con-ference on Computer Vision (ECCV), 2012.

[34] Tsung-Yi Lin, Michael Maire, Serge Belongie,James Hays, Pietro Perona, Deva Ramanan, Pi-otr Dollar, and C.Lawrence Zitnick. Microsoft

Page 20: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

COCO: Common objects in context. In Euro-pean Conference on Computer Vision (ECCV),pages 740–755, 2014.

[35] Joan M. Llargues, Juan Peralta, Raul Arrabales,Manuel Gonzalez, Paulo Cortez, and Antonio M.Lopez. Artificial intelligence approaches for thegeneration and assessment of believable human-like behaviour in virtual characters. ExpertSystems With Applications, 41(16):7281–7290,2014.

[36] Javier Marın, David Vazquez, David Geronimo,and Antonio M. Lopez. Learning appearancein virtual scenarios for pedestrian detection. InIEEE Conference on Computer Vision and Pat-tern Recognition (CVPR), 2010.

[37] Francisco Massa, Bryan C. Russell, and Math-ieu Aubry. Deep exemplar 2D-3D detection byadapting from real to rendered views. In IEEEConference on Computer Vision and PatternRecognition (CVPR), 2016.

[38] Nikolaus Mayer, Eddy Ilg, Philip Hausser,Philipp Fischer, Daniel Cremers, Alexey Doso-vitskiy, and Thomas Brox. A large dataset totrain convolutional networks for disparity, opti-cal flow, and scene flow estimation. In IEEEConference on Computer Vision and PatternRecognition (CVPR), 2016.

[39] Stephan Meister and Daniel Kondermann. Realversus realistically rendered scenes for opticalflow evaluation. In Conference on Electronic Me-dia Technology (CEMT), 2011.

[40] Volodymyr Mnih, Koray Kavukcuoglu, DavidSilver, Alex Graves, Ioannis Antonoglou, DaanWierstra, and Martin Riedmiller. Playing Atariwith deep reinforcement learning. In AnnualConference on Neural Information ProcessingSystems (NIPS), Workshop on Deep Learning,2013.

[41] Yair Movshovitz-Attias, Takeo Kanade, andYaser Sheikh. How useful is photo-realisticrendering for visual learning? CoRR,arXiv:1603.08152, 2016.

[42] Naveen Onkarappa and Angel D. Sappa. Syn-thetic sequences and ground-truth flow field gen-eration for algorithm validation. MultimediaTools and Applications, 74(9):3121–3135, 2015.

[43] Pau Panareda-Busto, Joerg Liebelt, and JuergenGall. Adaptation of synthetic data for coarse-to-fine viewpoint refinement. In BMVA BritishMachine Vision Conference (BMVC), 2015.

[44] Jeremie Papon and Markus Schoeler. Semanticpose using deep networks trained on syntheticRGB-D. In IEEE International Conference onComputer Vision (ICCV), 2015.

[45] Xingchao Peng, Baochen Sun, Karim Ali, andKate Saenko. Learning deep object detectorsfrom 3D models. In IEEE International Con-ference on Computer Vision (ICCV), 2015.

[46] Bojan Pepik, Michael Stark, Peter Gehler, andBernt Schiele. Teaching 3D geometry to de-formable part models. In IEEE Conferenceon Computer Vision and Pattern Recognition(CVPR), 2012.

[47] Leonid Pishchulin, Arjun Jain, Christian Wojek,Mykhaylo Andriluka, Thorsten Thormahlen,and Bernt Schiele. Learning people detectionmodels from few training samples. In IEEE Con-ference on Computer Vision and Pattern Recog-nition (CVPR), 2011.

[48] Amazon Mechanical Turk.http://www.mturk.com.

[49] Stephan R. Richter, Vibhav Vineet, StefanRoth, and Koltun Vladlen. Playing for data:Ground truth from computer games. In Euro-pean Conference on Computer Vision (ECCV),2016.

[50] German Ros, Laura Sellart, Joanna Materzyska,David Vazquez, and Antonio M. Lopez. TheSYNTHIA dataset: a large collection of syn-thetic images for semantic segmentation of ur-ban scenes. In IEEE Conference on ComputerVision and Pattern Recognition (CVPR), 2016.

Page 21: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

[51] German Ros, Simon Stent, Pablo F. Al-cantarilla, and Tomoki Watanabe. Train-ing constrained deconvolutional networks forroad scene semantic segmentation. CoRR,arXiv:1603.08152, 2016.

[52] Artem Rozantsev, Vincent Lepetit, and PascalFua. On rendering synthetic images for trainingan object detector. Computer Vision and ImageUnderstanding, 137:24–37, 2015.

[53] Bryan C. Russell, Antonio Torralba, Kevin P.Murphy, and William T. Freeman. LabelMe: adatabase and web-based tootl for image annota-tion. International Journal of Computer Vision(IJCV), 77(1–3):157–173, 2008.

[54] Jorge Sanchez, Florent Perronnin, ThomasMensink, and Jakob Verbeek. Image classifica-tion with the fisher vector: Theory and prac-tice. International Journal of Computer Vision(IJCV), 105(3):222–245, 2013.

[55] Scott Satkin, Michael Goesele, and BerntSchiele. Back to the future: Learning shapemodels from 3D CAD data. In BMVA BritishMachine Vision Conference (BMVC), 2010.

[56] Scott Satkin, Jason Lin, and Martial Hebert.Data-driven scene understanding from 3D mod-els. In BMVA British Machine Vision Confer-ence (BMVC), 2012.

[57] Johannes Schels, Jorg Liebelt, Klaus Schertler,and Rainer Lienhart. Synthetically trainedmulti-view object class and viewpoint detectionfor advanced image retrieval. In InternationalConference on Multimedia Retrieval (ICMR),2011.

[58] Alireza Shafaei, James J. Little, and MarkSchmidt. Play and learn: Using video games totrain computer vision models. In BMVA BritishMachine Vision Conference (BMVC), 2016.

[59] Jamie Shotton, Andrew Fitzgibbon, Mat Cook,Toby Sharp, Mark Finocchio, Richard Moore,Alex Kipmanand, and Andrew Blake. Real-timehuman pose recognition in parts from a single

depth image. In IEEE Conference on ComputerVision and Pattern Recognition (CVPR), 2011.

[60] Yainuvis Socarras, Sebastian Ramos, DavidVazquez, Antonio M. Lopez, and Theo Gevers.Adapting pedestrian detection from syntheticto far infrared images. In IEEE InternationalConference on Computer Vision (ICCV), VisualDomain Adaptation and Dataset Bias (ICCV-VisDA), 2013.

[61] Hao Su, Charles R. Qi, Yangyan Yi, andLeonidas Guibas. Render for CNN: viewpointestimation in images using CNNs trained withrendered 3D model views. In IEEE InternationalConference on Computer Vision (ICCV), 2015.

[62] Hao Su, Fan Wang, Yangyan Yi, and LeonidasGuibas. 3D-assisted feature synthesis for novelviews of an object. In IEEE International Con-ference on Computer Vision (ICCV), 2015.

[63] Baochen Sun and Kate Saenko. From virtual toreality: Fast adaptation of virtual object detec-tors to real domains. In BMVA British MachineVision Conference (BMVC), 2014.

[64] Geoffrey R. Taylor, Andrew J. Chosak, and PaulC. Brewer. OVVV: Using virtual worlds to de-sign and evaluate surveillance systems. In IEEEConference on Computer Vision and PatternRecognition (CVPR), 2007.

[65] Tatiana Tommasi, Novi Patricia, Barbara Ca-puto, and Tinne Tuytelaars. A deeper look atdataset bias. In German Conference on PatternRecognition (GCPR), 2015.

[66] Antonio Torralba and Alexei A. Efros. Unbi-ased look at dataset bias. In IEEE Conferenceon Computer Vision and Pattern Recognition(CVPR), 2011.

[67] Eric Tzeng, Judy Hoffman, Trevor Darrell, andKate Saenko. Simultaneous deep transfer acrossdomains and tasks. In IEEE InternationalConference on Computer Vision (ICCV), pages4068–4076, 2015.

Page 22: arXiv:1612.09134v1 [cs.CV] 29 Dec 2016refbase.cvc.uab.es/files/LXG2016.pdf · December 30, 2016 Abstract ... arXiv:1612.09134v1 [cs.CV] 29 Dec 2016. Invited book chapter to appear

Invited book chapter to appear in Domain Adaptation in Computer Vision Applications, Springer Series: Advances

in Computer Vision and Pattern Recognition, Edited by Gabriela Csurka. Written during Summer 2016.

[68] David Vazquez, Antonio M. Lopez, JavierMarın, Daniel Ponsa, and David Geronimo. Vir-tual and real world adaptation for pedestriandetection. Transactions of Pattern Recognitionand Machine Analyses (PAMI), 36(4):797 – 809,2014.

[69] David Vazquez, Antonio M. Lopez, DanielPonsa, and David Geronimo. Interactive train-ing of human detectors. In Angel D. Sappa andJordi Vitria, editors, Multimodal Interaction inImage and Video Applications Intelligent Sys-tems, pages 169–184. Springer, 2013.

[70] David Vazquez, Antonio M. Lopez, DanielPonsa, and Javier Marın. Cool world: domainadaptation of virtual and real worlds for humandetection using active learning. In Annual Con-ference on Neural Information Processing Sys-tems (NIPS), Workshop on Domain Adaptation:Theory and Applications, 2011.

[71] David Vazquez, Antonio M. Lopez, DanielPonsa, and Javier Marın. Virtual worlds andactive learning for human detection. In Inter-national Conference on Multimodal Interaction(ICMI), 2011.

[72] Ramakrishna Vedantam, Xiao Lin, Tanmay Ba-tra, C. Lawrence Zitnick, and Devi Parikh.Learning common sense through visual abstrac-tion. In IEEE International Conference on Com-puter Vision (ICCV), 2015.

[73] V.S.R. Veeravasarapu, Rudra Narayan Hota,Constantin Rothkopf, and Ramesh Visvanathan.Model validation for vision systems via graphicssimulation. CoRR, arXiv:1512.01401, 2015.

[74] V.S.R. Veeravasarapu, Rudra Narayan Hota,Constantin Rothkopf, and Ramesh Visvanathan.Simulations for validation of vision systems.CoRR, arXiv:1512.01030, 2015.

[75] V.S.R. Veeravasarapu, Constantin Rothkopf,and Ramesh Visvanathan. Model-driven simu-lations for deep convolutional neural networks.CoRR, arXiv:1605.09582, 2016.

[76] Paul Viola and Michael Jones. Rapid object de-tection using a boosted cascade of simple fea-tures. In IEEE Conference on Computer Visionand Pattern Recognition (CVPR), 2001.

[77] Jiaolong Xu, Sebastian Ramos, David Vazquez,and Antonio M. Lopez. Domain adaptationof deformable part-based models. Transactionsof Pattern Recognition and Machine Analyses(PAMI), 36(12):2367–2380, 2014.

[78] Jiaolong Xu, Sebastian Ramos, David Vazquez,and Antonio M. Lopez. Hierarchical adaptivestructural SVM for domain adaptation. Inter-national Journal of Computer Vision (IJCV),119(2):159–178, 2016.

[79] Jiaolong Xu, David Vazquez, Antonio M. Lopez,Javier Marın, and Daniel Ponsa. Learninga part-based pedestrian detector in a virtualworld. Transactions on Intelligent Transporta-tion Systems (ITS), 15(5):2121–2131, 2014.

[80] Jiaolong Xu, David Vazquez, Krystian Mikola-jczyk, and Antonio M. Lopez. Hierarchical on-line domain adaptation of deformable part-basedmodels. In International Conference on Roboticsand Automation (ICRA), 2016.

[81] Xiangxin Zhu, Carl Vondrick, Charless C.Fowlkes, and Deva Ramanan. Do we need moretraining data? International Journal of Com-puter Vision (IJCV), 119(1):76–92, 2016.

[82] Yuke Zhu, Roozbeh Mottaghi, Eric Kolve,Joseph J. Lim, and Abhinav Gupta. Target-driven visual navigation in indoor scenes us-ing deep reinforcement learning. CoRR,arXiv:1609.05143, 2016.

[83] C. Lawrence Zitnick, Ramakrishna Vedantam,and Devi Parikh. Adopting abstract imagesfor semantic scene understanding. Transactionsof Pattern Recognition and Machine Analyses(PAMI), 38(4):627–638, 2016.