104,840 research outputs found
In Defense of the Classification Loss for Person Re-Identification
The recent research for person re-identification has been focused on two
trends. One is learning the part-based local features to form more informative
feature descriptors. The other is designing effective metric learning loss
functions such as the triplet loss family. We argue that learning global
features with classification loss could achieve the same goal, even with some
simple and cost-effective architecture design. In this paper, we first explain
why the person re-id framework with standard classification loss usually has
inferior performance compared to metric learning. Based on that, we further
propose a person re-id framework featured by channel grouping and multi-branch
strategy, which divides global features into multiple channel groups and learns
the discriminative channel group features by multi-branch classification
layers. The extensive experiments show that our framework outperforms prior
state-of-the-arts in terms of both accuracy and inference speed
Metric Attack and Defense for Person Re-identification
Person re-identification (re-ID) has attracted much attention recently due to
its great importance in video surveillance. In general, distance metrics used
to identify two person images are expected to be robust under various
appearance changes. However, our work observes the extreme vulnerability of
existing distance metrics to adversarial examples, generated by simply adding
human-imperceptible perturbations to person images. Hence, the security danger
is dramatically increased when deploying commercial re-ID systems in video
surveillance.
Although adversarial examples have been extensively applied for
classification analysis, it is rarely studied in metric analysis like person
re-identification. The most likely reason is the natural gap between the
training and testing of re-ID networks, that is, the predictions of a re-ID
network cannot be directly used during testing without an effective metric. In
this work, we bridge the gap by proposing Adversarial Metric Attack, a parallel
methodology to adversarial classification attacks. Comprehensive experiments
clearly reveal the adversarial effects in re-ID systems. Meanwhile, we also
present an early attempt of training a metric-preserving network, thereby
defending the metric against adversarial attacks. At last, by benchmarking
various adversarial settings, we expect that our work can facilitate the
development of adversarial attack and defense in metric-based applications
Cross-Resolution Person Re-identification with Deep Antithetical Learning
Images with different resolutions are ubiquitous in public person
re-identification (ReID) datasets and real-world scenes, it is thus crucial for
a person ReID model to handle the image resolution variations for improving its
generalization ability. However, most existing person ReID methods pay little
attention to this resolution discrepancy problem. One paradigm to deal with
this problem is to use some complicated methods for mapping all images into an
artificial image space, which however will disrupt the natural image
distribution and requires heavy image preprocessing. In this paper, we analyze
the deficiencies of several widely-used objective functions handling image
resolution discrepancies and propose a new framework called deep antithetical
learning that directly learns from the natural image space rather than creating
an arbitrary one. We first quantify and categorize original training images
according to their resolutions. Then we create an antithetical training set and
make sure that original training images have counterparts with antithetical
resolutions in this new set. At last, a novel Contrastive Center Loss(CCL) is
proposed to learn from images with different resolutions without being
interfered by their resolution discrepancies. Extensive experimental analyses
and evaluations indicate that the proposed framework, even using a vanilla deep
ReID network, exhibits remarkable performance improvements. Without bells and
whistles, our approach outperforms previous state-of-the-art methods by a large
margin
Metric Embedding Autoencoders for Unsupervised Cross-Dataset Transfer Learning
Cross-dataset transfer learning is an important problem in person
re-identification (Re-ID). Unfortunately, not too many deep transfer Re-ID
models exist for realistic settings of practical Re-ID systems. We propose a
purely deep transfer Re-ID model consisting of a deep convolutional neural
network and an autoencoder. The latent code is divided into metric embedding
and nuisance variables. We then utilize an unsupervised training method that
does not rely on co-training with non-deep models. Our experiments show
improvements over both the baseline and competitors' transfer learning models.Comment: ICANN 2018 (The 27th International Conference on Artificial Neural
Networks) proceedin
AlignedReID: Surpassing Human-Level Performance in Person Re-Identification
In this paper, we propose a novel method called AlignedReID that extracts a
global feature which is jointly learned with local features. Global feature
learning benefits greatly from local feature learning, which performs an
alignment/matching by calculating the shortest path between two sets of local
features, without requiring extra supervision. After the joint learning, we
only keep the global feature to compute the similarities between images. Our
method achieves rank-1 accuracy of 94.4% on Market1501 and 97.8% on CUHK03,
outperforming state-of-the-art methods by a large margin. We also evaluate
human-level performance and demonstrate that our method is the first to surpass
human-level performance on Market1501 and CUHK03, two widely used Person ReID
datasets.Comment: 9 pages, 8 figure
Adaptation and Re-Identification Network: An Unsupervised Deep Transfer Learning Approach to Person Re-Identification
Person re-identification (Re-ID) aims at recognizing the same person from
images taken across different cameras. To address this task, one typically
requires a large amount labeled data for training an effective Re-ID model,
which might not be practical for real-world applications. To alleviate this
limitation, we choose to exploit a sufficient amount of pre-existing labeled
data from a different (auxiliary) dataset. By jointly considering such an
auxiliary dataset and the dataset of interest (but without label information),
our proposed adaptation and re-identification network (ARN) performs
unsupervised domain adaptation, which leverages information across datasets and
derives domain-invariant features for Re-ID purposes. In our experiments, we
verify that our network performs favorably against state-of-the-art
unsupervised Re-ID approaches, and even outperforms a number of baseline Re-ID
methods which require fully supervised data for training.Comment: 7 pages, 3 figures. CVPR 2018 workshop pape
Virtual CNN Branching: Efficient Feature Ensemble for Person Re-Identification
In this paper we introduce an ensemble method for convolutional neural
network (CNN), called "virtual branching," which can be implemented with nearly
no additional parameters and computation on top of standard CNNs. We propose
our method in the context of person re-identification (re-ID). Our CNN model
consists of shared bottom layers, followed by "virtual" branches, where neurons
from a block of regular convolutional and fully-connected layers are
partitioned into multiple sets. Each virtual branch is trained with different
data to specialize in different aspects, e.g., a specific body region or pose
orientation. In this way, robust ensemble representations are obtained against
human body misalignment, deformations, or variations in viewing angles, at
nearly no any additional cost. The proposed method achieves competitive
performance on multiple person re-ID benchmark datasets, including Market-1501,
CUHK03, and DukeMTMC-reID
Imitating Targets from all sides: An Unsupervised Transfer Learning method for Person Re-identification
Person re-identification (Re-ID) models usually show a limited performance
when they are trained on one dataset and tested on another dataset due to the
inter-dataset bias (e.g. completely different identities and backgrounds) and
the intra-dataset difference (e.g. camera invariance). In terms of this issue,
given a labelled source training set and an unlabelled target training set, we
propose an unsupervised transfer learning method characterized by 1) bridging
inter-dataset bias and intra-dataset difference via a proposed ImitateModel
simultaneously; 2) regarding the unsupervised person Re-ID problem as a
semi-supervised learning problem formulated by a dual classification loss to
learn a discriminative representation across domains; 3) exploiting the
underlying commonality across different domains from the class-style space to
improve the generalization ability of re-ID models. Extensive experiments are
conducted on two widely employed benchmarks, including Market-1501 and
DukeMTMC-reID, and experimental results demonstrate that the proposed method
can achieve a competitive performance against other state-of-the-art
unsupervised Re-ID approaches
VOC-ReID: Vehicle Re-identification based on Vehicle-Orientation-Camera
Vehicle re-identification is a challenging task due to high intra-class
variances and small inter-class variances. In this work, we focus on the
failure cases caused by similar background and shape. They pose serve bias on
similarity, making it easier to neglect fine-grained information. To reduce the
bias, we propose an approach named VOC-ReID, taking the triplet
vehicle-orientation-camera as a whole and reforming background/shape similarity
as camera/orientation re-identification. At first, we train models for vehicle,
orientation and camera re-identification respectively. Then we use orientation
and camera similarity as penalty to get final similarity. Besides, we propose a
high performance baseline boosted by bag of tricks and weakly supervised data
augmentation. Our algorithm achieves the second place in vehicle
re-identification at the NVIDIA AI City Challenge 2020.Comment: AICity2020 Challenge, CVPR 2020 workshop, code avaible at github(link
in abstract
Unsupervised Person Re-identification by Deep Learning Tracklet Association
Mostexistingpersonre-identification(re-id)methods relyon supervised model
learning on per-camera-pair manually labelled pairwise training data. This
leads to poor scalability in practical re-id deployment due to the lack of
exhaustive identity labelling of image positive and negative pairs for every
camera pair. In this work, we address this problem by proposing an unsupervised
re-id deep learning approach capable of incrementally discovering and
exploiting the underlying re-id discriminative information from automatically
generated person tracklet data from videos in an end-to-end model optimisation.
We formulate a Tracklet Association Unsupervised Deep Learning (TAUDL)
framework characterised by jointly learning per-camera (within-camera) tracklet
association (labelling) and cross-camera tracklet correlation by maximising the
discovery of most likely tracklet relationships across camera views. Extensive
experiments demonstrate the superiority of the proposed TAUDL model over the
state-of-the-art unsupervised and domain adaptation re- id methods using six
person re-id benchmarking datasets.Comment: ECCV 2018 Ora
- …