330 research outputs found
Unsupervised Person Re-identification by Soft Multilabel Learning
Although unsupervised person re-identification (RE-ID) has drawn increasing
research attentions due to its potential to address the scalability problem of
supervised RE-ID models, it is very challenging to learn discriminative
information in the absence of pairwise labels across disjoint camera views. To
overcome this problem, we propose a deep model for the soft multilabel learning
for unsupervised RE-ID. The idea is to learn a soft multilabel (real-valued
label likelihood vector) for each unlabeled person by comparing (and
representing) the unlabeled person with a set of known reference persons from
an auxiliary domain. We propose the soft multilabel-guided hard negative mining
to learn a discriminative embedding for the unlabeled target domain by
exploring the similarity consistency of the visual features and the soft
multilabels of unlabeled target pairs. Since most target pairs are cross-view
pairs, we develop the cross-view consistent soft multilabel learning to achieve
the learning goal that the soft multilabels are consistently good across
different camera views. To enable effecient soft multilabel learning, we
introduce the reference agent learning to represent each reference person by a
reference agent in a joint embedding. We evaluate our unified deep model on
Market-1501 and DukeMTMC-reID. Our model outperforms the state-of-the-art
unsupervised RE-ID methods by clear margins. Code is available at
https://github.com/KovenYu/MAR.Comment: CVPR19, ora
Exploiting Sample Uncertainty for Domain Adaptive Person Re-Identification
Many unsupervised domain adaptive (UDA) person re-identification (ReID)
approaches combine clustering-based pseudo-label prediction with feature
fine-tuning. However, because of domain gap, the pseudo-labels are not always
reliable and there are noisy/incorrect labels. This would mislead the feature
representation learning and deteriorate the performance. In this paper, we
propose to estimate and exploit the credibility of the assigned pseudo-label of
each sample to alleviate the influence of noisy labels, by suppressing the
contribution of noisy samples. We build our baseline framework using the mean
teacher method together with an additional contrastive loss. We have observed
that a sample with a wrong pseudo-label through clustering in general has a
weaker consistency between the output of the mean teacher model and the student
model. Based on this finding, we propose to exploit the uncertainty (measured
by consistency levels) to evaluate the reliability of the pseudo-label of a
sample and incorporate the uncertainty to re-weight its contribution within
various ReID losses, including the identity (ID) classification loss per
sample, the triplet loss, and the contrastive loss. Our uncertainty-guided
optimization brings significant improvement and achieves the state-of-the-art
performance on benchmark datasets.Comment: 9 pages. Accepted to 35th AAAI Conference on Artificial Intelligence
(AAAI 2021
Unsupervised Disentanglement GAN for Domain Adaptive Person Re-Identification
While recent person re-identification (ReID) methods achieve high accuracy in
a supervised setting, their generalization to an unlabelled domain is still an
open problem. In this paper, we introduce a novel unsupervised disentanglement
generative adversarial network (UD-GAN) to address the domain adaptation issue
of supervised person ReID. Our framework jointly trains a ReID network for
discriminative features extraction in a source labelled domain using identity
annotation, and adapts the ReID model to an unlabelled target domain by
learning disentangled latent representations on the domain. Identity-unrelated
features in the target domain are distilled from the latent features. As a
result, the ReID features better encompass the identity of a person in the
unsupervised domain. We conducted experiments on the Market1501, DukeMTMC and
MSMT17 datasets. Results show that the unsupervised domain adaptation problem
in ReID is very challenging. Nevertheless, our method shows improvement in half
of the domain transfers and achieve state-of-the-art performance for one of
them.Comment: 8 pages, 5 figures, submitted to ICPR 202
Interpretable and Generalizable Person Re-Identification with Query-Adaptive Convolution and Temporal Lifting
For person re-identification, existing deep networks often focus on
representation learning. However, without transfer learning, the learned model
is fixed as is, which is not adaptable for handling various unseen scenarios.
In this paper, beyond representation learning, we consider how to formulate
person image matching directly in deep feature maps. We treat image matching as
finding local correspondences in feature maps, and construct query-adaptive
convolution kernels on the fly to achieve local matching. In this way, the
matching process and results are interpretable, and this explicit matching is
more generalizable than representation features to unseen scenarios, such as
unknown misalignments, pose or viewpoint changes. To facilitate end-to-end
training of this architecture, we further build a class memory module to cache
feature maps of the most recent samples of each class, so as to compute image
matching losses for metric learning. Through direct cross-dataset evaluation,
the proposed Query-Adaptive Convolution (QAConv) method gains large
improvements over popular learning methods (about 10%+ mAP), and achieves
comparable results to many transfer learning methods. Besides, a model-free
temporal cooccurrence based score weighting method called TLift is proposed,
which improves the performance to a further extent, achieving state-of-the-art
results in cross-dataset person re-identification. Code is available at
https://github.com/ShengcaiLiao/QAConv.Comment: This is the ECCV 2020 version, including the appendi
- …