8 research outputs found
Unsupervised Scene Adaptation with Memory Regularization in vivo
We consider the unsupervised scene adaptation problem of learning from both
labeled source data and unlabeled target data. Existing methods focus on
minoring the inter-domain gap between the source and target domains. However,
the intra-domain knowledge and inherent uncertainty learned by the network are
under-explored. In this paper, we propose an orthogonal method, called memory
regularization in vivo to exploit the intra-domain knowledge and regularize the
model training. Specifically, we refer to the segmentation model itself as the
memory module, and minor the discrepancy of the two classifiers, i.e., the
primary classifier and the auxiliary classifier, to reduce the prediction
inconsistency. Without extra parameters, the proposed method is complementary
to the most existing domain adaptation methods and could generally improve the
performance of existing methods. Albeit simple, we verify the effectiveness of
memory regularization on two synthetic-to-real benchmarks: GTA5 -> Cityscapes
and SYNTHIA -> Cityscapes, yielding +11.1% and +11.3% mIoU improvement over the
baseline model, respectively. Besides, a similar +12.0% mIoU improvement is
observed on the cross-city benchmark: Cityscapes -> Oxford RobotCar.Comment: 7 pages, 4 figures, 6 table
Unsupervised scene adaptation with memory regularization in vivo
This work focuses on the unsupervised scene adaptation problem of learning from both labeled source data and unlabeled target data. Existing approaches focus on minoring the inter-domain gap between the source and target domains. However, the intra-domain knowledge and inherent uncertainty learned by the network are under-explored. In this paper, we propose an orthogonal method, called memory regularization in vivo, to exploit the intra-domain knowledge and regularize the model training. Specifically, we refer to the segmentation model itself as the memory module, and minor the discrepancy of the two classifiers, i.e., the primary classifier and the auxiliary classifier, to reduce the prediction inconsistency. Without extra parameters, the proposed method is complementary to most existing domain adaptation methods and could generally improve the performance of existing methods. Albeit simple, we verify the effectiveness of memory regularization on two synthetic-to-real benchmarks: GTA5 ? Cityscapes and SYNTHIA ? Cityscapes, yielding +11.1% and +11.3% mIoU improvement over the baseline model, respectively. Besides, a similar +12.0% mIoU improvement is observed on the cross-city benchmark: Cityscapes ? Oxford RobotCar
DACS: Domain Adaptation via Cross-domain Mixed Sampling
Semantic segmentation models based on convolutional neural networks have
recently displayed remarkable performance for a multitude of applications.
However, these models typically do not generalize well when applied on new
domains, especially when going from synthetic to real data. In this paper we
address the problem of unsupervised domain adaptation (UDA), which attempts to
train on labelled data from one domain (source domain), and simultaneously
learn from unlabelled data in the domain of interest (target domain). Existing
methods have seen success by training on pseudo-labels for these unlabelled
images. Multiple techniques have been proposed to mitigate low-quality
pseudo-labels arising from the domain shift, with varying degrees of success.
We propose DACS: Domain Adaptation via Cross-domain mixed Sampling, which mixes
images from the two domains along with the corresponding labels and
pseudo-labels. These mixed samples are then trained on, in addition to the
labelled data itself. We demonstrate the effectiveness of our solution by
achieving state-of-the-art results for GTA5 to Cityscapes, a common
synthetic-to-real semantic segmentation benchmark for UDA.Comment: This paper has been accepted to WACV202
Adaptive Boosting for Domain Adaptation: Towards Robust Predictions in Scene Segmentation
Domain adaptation is to transfer the shared knowledge learned from the source
domain to a new environment, i.e., target domain. One common practice is to
train the model on both labeled source-domain data and unlabeled target-domain
data. Yet the learned models are usually biased due to the strong supervision
of the source domain. Most researchers adopt the early-stopping strategy to
prevent over-fitting, but when to stop training remains a challenging problem
since the lack of the target-domain validation set. In this paper, we propose
one efficient bootstrapping method, called Adaboost Student, explicitly
learning complementary models during training and liberating users from
empirical early stopping. Adaboost Student combines the deep model learning
with the conventional training strategy, i.e., adaptive boosting, and enables
interactions between learned models and the data sampler. We adopt one adaptive
data sampler to progressively facilitate learning on hard samples and aggregate
"weak" models to prevent over-fitting. Extensive experiments show that (1)
Without the need to worry about the stopping time, AdaBoost Student provides
one robust solution by efficient complementary model learning during training.
(2) AdaBoost Student is orthogonal to most domain adaptation methods, which can
be combined with existing approaches to further improve the state-of-the-art
performance. We have achieved competitive results on three widely-used scene
segmentation domain adaptation benchmarks.Comment: 10 pages, 7 tables, 5 figure