143,986 research outputs found
Exploiting feature representations through similarity learning, post-ranking and ranking aggregation for person re-identification
Person re-identification has received special attention by the human analysis
community in the last few years. To address the challenges in this field, many
researchers have proposed different strategies, which basically exploit either
cross-view invariant features or cross-view robust metrics. In this work, we
propose to exploit a post-ranking approach and combine different feature
representations through ranking aggregation. Spatial information, which
potentially benefits the person matching, is represented using a 2D body model,
from which color and texture information are extracted and combined. We also
consider background/foreground information, automatically extracted via Deep
Decompositional Network, and the usage of Convolutional Neural Network (CNN)
features. To describe the matching between images we use the polynomial feature
map, also taking into account local and global information. The Discriminant
Context Information Analysis based post-ranking approach is used to improve
initial ranking lists. Finally, the Stuart ranking aggregation method is
employed to combine complementary ranking lists obtained from different feature
representations. Experimental results demonstrated that we improve the
state-of-the-art on VIPeR and PRID450s datasets, achieving 67.21% and 75.64% on
top-1 rank recognition rate, respectively, as well as obtaining competitive
results on CUHK01 dataset.Comment: Preprint submitted to Image and Vision Computin
Learning Deep Context-aware Features over Body and Latent Parts for Person Re-identification
Person Re-identification (ReID) is to identify the same person across
different cameras. It is a challenging task due to the large variations in
person pose, occlusion, background clutter, etc How to extract powerful
features is a fundamental problem in ReID and is still an open problem today.
In this paper, we design a Multi-Scale Context-Aware Network (MSCAN) to learn
powerful features over full body and body parts, which can well capture the
local context knowledge by stacking multi-scale convolutions in each layer.
Moreover, instead of using predefined rigid parts, we propose to learn and
localize deformable pedestrian parts using Spatial Transformer Networks (STN)
with novel spatial constraints. The learned body parts can release some
difficulties, eg pose variations and background clutters, in part-based
representation. Finally, we integrate the representation learning processes of
full body and body parts into a unified framework for person ReID through
multi-class person identification tasks. Extensive evaluations on current
challenging large-scale person ReID datasets, including the image-based
Market1501, CUHK03 and sequence-based MARS datasets, show that the proposed
method achieves the state-of-the-art results.Comment: Accepted by CVPR 201
Context-aware person identification in personal photo collections
Identifying the people in photos is an important need for users of photo management systems. We present MediAssist, one such system which facilitates browsing, searching and semi-automatic annotation of personal photos, using analysis of both image content and the context in which the photo is captured. This semi-automatic annotation includes annotation of the identity of people in photos. In this paper, we focus on such person annotation, and propose person identification techniques based on a combination of context and content. We propose language modelling and nearest neighbor approaches to context-based person identification, in addition to novel face color and image color content-based features (used alongside face recognition and body patch features). We conduct a comprehensive empirical study of these techniques using the real private photo collections of a number of users, and show that combining context- and content-based analysis improves performance over content or context alone
Pedestrian Attribute Recognition: A Survey
Recognizing pedestrian attributes is an important task in computer vision
community due to it plays an important role in video surveillance. Many
algorithms has been proposed to handle this task. The goal of this paper is to
review existing works using traditional methods or based on deep learning
networks. Firstly, we introduce the background of pedestrian attributes
recognition (PAR, for short), including the fundamental concepts of pedestrian
attributes and corresponding challenges. Secondly, we introduce existing
benchmarks, including popular datasets and evaluation criterion. Thirdly, we
analyse the concept of multi-task learning and multi-label learning, and also
explain the relations between these two learning algorithms and pedestrian
attribute recognition. We also review some popular network architectures which
have widely applied in the deep learning community. Fourthly, we analyse
popular solutions for this task, such as attributes group, part-based,
\emph{etc}. Fifthly, we shown some applications which takes pedestrian
attributes into consideration and achieve better performance. Finally, we
summarized this paper and give several possible research directions for
pedestrian attributes recognition. The project page of this paper can be found
from the following website:
\url{https://sites.google.com/view/ahu-pedestrianattributes/}.Comment: Check our project page for High Resolution version of this survey:
https://sites.google.com/view/ahu-pedestrianattributes
Query-guided End-to-End Person Search
Person search has recently gained attention as the novel task of finding a
person, provided as a cropped sample, from a gallery of non-cropped images,
whereby several other people are also visible. We believe that i. person
detection and re-identification should be pursued in a joint optimization
framework and that ii. the person search should leverage the query image
extensively (e.g. emphasizing unique query patterns). However, so far, no prior
art realizes this. We introduce a novel query-guided end-to-end person search
network (QEEPS) to address both aspects. We leverage a most recent joint
detector and re-identification work, OIM [37]. We extend this with i. a
query-guided Siamese squeeze-and-excitation network (QSSE-Net) that uses global
context from both the query and gallery images, ii. a query-guided region
proposal network (QRPN) to produce query-relevant proposals, and iii. a
query-guided similarity subnetwork (QSimNet), to learn a query-guided
reidentification score. QEEPS is the first end-to-end query-guided detection
and re-id network. On both the most recent CUHK-SYSU [37] and PRW [46]
datasets, we outperform the previous state-of-the-art by a large margin.Comment: Accepted as poster in CVPR 201
- …