2 research outputs found
Joint Semantic Domain Alignment and Target Classifier Learning for Unsupervised Domain Adaptation
Unsupervised domain adaptation aims to transfer the classifier learned from
the source domain to the target domain in an unsupervised manner. With the help
of target pseudo-labels, aligning class-level distributions and learning the
classifier in the target domain are two widely used objectives. Existing
methods often separately optimize these two individual objectives, which makes
them suffer from the neglect of the other. However, optimizing these two
aspects together is not trivial. To alleviate the above issues, we propose a
novel method that jointly optimizes semantic domain alignment and target
classifier learning in a holistic way. The joint optimization mechanism can not
only eliminate their weaknesses but also complement their strengths. The
theoretical analysis also verifies the favor of the joint optimization
mechanism. Extensive experiments on benchmark datasets show that the proposed
method yields the best performance in comparison with the state-of-the-art
unsupervised domain adaptation methods
Joint Contrastive Learning for Unsupervised Domain Adaptation
Enhancing feature transferability by matching marginal distributions has led
to improvements in domain adaptation, although this is at the expense of
feature discrimination. In particular, the ideal joint hypothesis error in the
target error upper bound, which was previously considered to be minute, has
been found to be significant, impairing its theoretical guarantee. In this
paper, we propose an alternative upper bound on the target error that
explicitly considers the joint error to render it more manageable. With the
theoretical analysis, we suggest a joint optimization framework that combines
the source and target domains. Further, we introduce Joint Contrastive Learning
(JCL) to find class-level discriminative features, which is essential for
minimizing the joint error. With a solid theoretical framework, JCL employs
contrastive loss to maximize the mutual information between a feature and its
label, which is equivalent to maximizing the Jensen-Shannon divergence between
conditional distributions. Experiments on two real-world datasets demonstrate
that JCL outperforms the state-of-the-art methods.Comment: 16 pages, 1 figure, 4 table