6,436 research outputs found
Maximum Classifier Discrepancy for Unsupervised Domain Adaptation
In this work, we present a method for unsupervised domain adaptation. Many
adversarial learning methods train domain classifier networks to distinguish
the features as either a source or target and train a feature generator network
to mimic the discriminator. Two problems exist with these methods. First, the
domain classifier only tries to distinguish the features as a source or target
and thus does not consider task-specific decision boundaries between classes.
Therefore, a trained generator can generate ambiguous features near class
boundaries. Second, these methods aim to completely match the feature
distributions between different domains, which is difficult because of each
domain's characteristics.
To solve these problems, we introduce a new approach that attempts to align
distributions of source and target by utilizing the task-specific decision
boundaries. We propose to maximize the discrepancy between two classifiers'
outputs to detect target samples that are far from the support of the source. A
feature generator learns to generate target features near the support to
minimize the discrepancy. Our method outperforms other methods on several
datasets of image classification and semantic segmentation. The codes are
available at \url{https://github.com/mil-tokyo/MCD_DA}Comment: Accepted to CVPR2018 Oral, Code is available at
https://github.com/mil-tokyo/MCD_D
Curriculum Domain Adaptation for Semantic Segmentation of Urban Scenes
During the last half decade, convolutional neural networks (CNNs) have
triumphed over semantic segmentation, which is one of the core tasks in many
applications such as autonomous driving. However, to train CNNs requires a
considerable amount of data, which is difficult to collect and laborious to
annotate. Recent advances in computer graphics make it possible to train CNNs
on photo-realistic synthetic imagery with computer-generated annotations.
Despite this, the domain mismatch between the real images and the synthetic
data cripples the models' performance. Hence, we propose a curriculum-style
learning approach to minimize the domain gap in urban scenery semantic
segmentation. The curriculum domain adaptation solves easy tasks first to infer
necessary properties about the target domain; in particular, the first task is
to learn global label distributions over images and local distributions over
landmark superpixels. These are easy to estimate because images of urban scenes
have strong idiosyncrasies (e.g., the size and spatial relations of buildings,
streets, cars, etc.). We then train a segmentation network while regularizing
its predictions in the target domain to follow those inferred properties. In
experiments, our method outperforms the baselines on two datasets and two
backbone networks. We also report extensive ablation studies about our
approach.Comment: This is the extended version of the ICCV 2017 paper "Curriculum
Domain Adaptation for Semantic Segmentation of Urban Scenes" with additional
GTA experimen
- …