1,261 research outputs found
Zero-Shot Deep Domain Adaptation
Domain adaptation is an important tool to transfer knowledge about a task
(e.g. classification) learned in a source domain to a second, or target domain.
Current approaches assume that task-relevant target-domain data is available
during training. We demonstrate how to perform domain adaptation when no such
task-relevant target-domain data is available. To tackle this issue, we propose
zero-shot deep domain adaptation (ZDDA), which uses privileged information from
task-irrelevant dual-domain pairs. ZDDA learns a source-domain representation
which is not only tailored for the task of interest but also close to the
target-domain representation. Therefore, the source-domain task of interest
solution (e.g. a classifier for classification tasks) which is jointly trained
with the source-domain representation can be applicable to both the source and
target representations. Using the MNIST, Fashion-MNIST, NIST, EMNIST, and SUN
RGB-D datasets, we show that ZDDA can perform domain adaptation in
classification tasks without access to task-relevant target-domain training
data. We also extend ZDDA to perform sensor fusion in the SUN RGB-D scene
classification task by simulating task-relevant target-domain representations
with task-relevant source-domain data. To the best of our knowledge, ZDDA is
the first domain adaptation and sensor fusion method which requires no
task-relevant target-domain data. The underlying principle is not particular to
computer vision data, but should be extensible to other domains.Comment: This paper is accepted to the European Conference on Computer Vision
(ECCV), 201
Residual Parameter Transfer for Deep Domain Adaptation
The goal of Deep Domain Adaptation is to make it possible to use Deep Nets
trained in one domain where there is enough annotated training data in another
where there is little or none. Most current approaches have focused on learning
feature representations that are invariant to the changes that occur when going
from one domain to the other, which means using the same network parameters in
both domains. While some recent algorithms explicitly model the changes by
adapting the network parameters, they either severely restrict the possible
domain changes, or significantly increase the number of model parameters.
By contrast, we introduce a network architecture that includes auxiliary
residual networks, which we train to predict the parameters in the domain with
little annotated data from those in the other one. This architecture enables us
to flexibly preserve the similarities between domains where they exist and
model the differences when necessary. We demonstrate that our approach yields
higher accuracy than state-of-the-art methods without undue complexity
- …