3 research outputs found
Global Distance-distributions Separation for Unsupervised Person Re-identification
Supervised person re-identification (ReID) often has poor scalability and
usability in real-world deployments due to domain gaps and the lack of
annotations for the target domain data. Unsupervised person ReID through domain
adaptation is attractive yet challenging. Existing unsupervised ReID approaches
often fail in correctly identifying the positive samples and negative samples
through the distance-based matching/ranking. The two distributions of distances
for positive sample pairs (Pos-distr) and negative sample pairs (Neg-distr) are
often not well separated, having large overlap. To address this problem, we
introduce a global distance-distributions separation (GDS) constraint over the
two distributions to encourage the clear separation of positive and negative
samples from a global view. We model the two global distance distributions as
Gaussian distributions and push apart the two distributions while encouraging
their sharpness in the unsupervised training process. Particularly, to model
the distributions from a global view and facilitate the timely updating of the
distributions and the GDS related losses, we leverage a momentum update
mechanism for building and maintaining the distribution parameters (mean and
variance) and calculate the loss on the fly during the training.
Distribution-based hard mining is proposed to further promote the separation of
the two distributions. We validate the effectiveness of the GDS constraint in
unsupervised ReID networks. Extensive experiments on multiple ReID benchmark
datasets show our method leads to significant improvement over the baselines
and achieves the state-of-the-art performance.Comment: Accepted by ECCV202
Large Margin Learning in Set to Set Similarity Comparison for Person Re-identification
Person re-identification (Re-ID) aims at matching images of the same person
across disjoint camera views, which is a challenging problem in multimedia
analysis, multimedia editing and content-based media retrieval communities. The
major challenge lies in how to preserve similarity of the same person across
video footages with large appearance variations, while discriminating different
individuals. To address this problem, conventional methods usually consider the
pairwise similarity between persons by only measuring the point to point (P2P)
distance. In this paper, we propose to use deep learning technique to model a
novel set to set (S2S) distance, in which the underline objective focuses on
preserving the compactness of intra-class samples for each camera view, while
maximizing the margin between the intra-class set and inter-class set. The S2S
distance metric is consisted of three terms, namely the class-identity term,
the relative distance term and the regularization term. The class-identity term
keeps the intra-class samples within each camera view gathering together, the
relative distance term maximizes the distance between the intra-class class set
and inter-class set across different camera views, and the regularization term
smoothness the parameters of deep convolutional neural network (CNN). As a
result, the final learned deep model can effectively find out the matched
target to the probe object among various candidates in the video gallery by
learning discriminative and stable feature representations. Using the CUHK01,
CUHK03, PRID2011 and Market1501 benchmark datasets, we extensively conducted
comparative evaluations to demonstrate the advantages of our method over the
state-of-the-art approaches.Comment: Accepted by IEEE Transactions on Multimedi
Deep Ranking Model by Large Adaptive Margin Learning for Person Re-identification
Person re-identification aims to match images of the same person across
disjoint camera views, which is a challenging problem in video surveillance.
The major challenge of this task lies in how to preserve the similarity of the
same person against large variations caused by complex backgrounds, mutual
occlusions and different illuminations, while discriminating the different
individuals. In this paper, we present a novel deep ranking model with feature
learning and fusion by learning a large adaptive margin between the intra-class
distance and inter-class distance to solve the person re-identification
problem. Specifically, we organize the training images into a batch of pairwise
samples. Treating these pairwise samples as inputs, we build a novel part-based
deep convolutional neural network (CNN) to learn the layered feature
representations by preserving a large adaptive margin. As a result, the final
learned model can effectively find out the matched target to the anchor image
among a number of candidates in the gallery image set by learning
discriminative and stable feature representations. Overcoming the weaknesses of
conventional fixed-margin loss functions, our adaptive margin loss function is
more appropriate for the dynamic feature space. On four benchmark datasets,
PRID2011, Market1501, CUHK01 and 3DPeS, we extensively conduct comparative
evaluations to demonstrate the advantages of the proposed method over the
state-of-the-art approaches in person re-identification.Comment: Accepted to Pattern Recognitio