22 research outputs found
Covariate conscious approach for Gait recognition based upon Zernike moment invariants
Gait recognition i.e. identification of an individual from his/her walking
pattern is an emerging field. While existing gait recognition techniques
perform satisfactorily in normal walking conditions, there performance tend to
suffer drastically with variations in clothing and carrying conditions. In this
work, we propose a novel covariate cognizant framework to deal with the
presence of such covariates. We describe gait motion by forming a single 2D
spatio-temporal template from video sequence, called Average Energy Silhouette
image (AESI). Zernike moment invariants (ZMIs) are then computed to screen the
parts of AESI infected with covariates. Following this, features are extracted
from Spatial Distribution of Oriented Gradients (SDOGs) and novel Mean of
Directional Pixels (MDPs) methods. The obtained features are fused together to
form the final well-endowed feature set. Experimental evaluation of the
proposed framework on three publicly available datasets i.e. CASIA dataset B,
OU-ISIR Treadmill dataset B and USF Human-ID challenge dataset with recently
published gait recognition approaches, prove its superior performance.Comment: 11 page
2.5D multi-view gait recognition based on point cloud registration
This paper presents a method for modeling a 2.5-dimensional (2.5D) human body and extracting the gait features for identifying the human subject. To achieve view-invariant gait recognition, a multi-view synthesizing method based on point cloud registration (MVSM) to generate multi-view training galleries is proposed. The concept of a density and curvature-based Color Gait Curvature Image is introduced to map 2.5D data onto a 2D space to enable data dimension reduction by discrete cosine transform and 2D principle component analysis. Gait recognition is achieved via a 2.5D view-invariant gait recognition method based on point cloud registration. Experimental results on the in-house database captured by a Microsoft Kinect camera show a significant performance gain when using MVSM
Video from nearly still: An application to low frame-rate gait recognition
In this paper, we propose a temporal super resolution ap-proach for quasi-periodic image sequence such as human gait. The proposed method effectively combines example-based and reconstruction-based temporal super resolution approaches. A periodic image sequence is expressed as a manifold parameterized by a phase and a standard mani-fold is learned from multiple high frame-rate sequences in the training stage. In the test stage, an initial phase for each frame of an input low frame-rate image sequence is estimated based on the standard manifold at first, and the manifold reconstruction and the phase estimation are then iterated to generate better high frame-rate images in the energy minimization framework that ensures the fitness to both the input images and the standard manifold. The pro-posed method is applied to low frame-rate gait recognition and experiments with real data of 100 subjects demonstrate a significant improvement by the proposed method, particu-larly for quite low frame-rate videos (e.g., 1 fps). 1
Gait recognition and understanding based on hierarchical temporal memory using 3D gait semantic folding
Gait recognition and understanding systems have shown a wide-ranging application prospect. However, their use of unstructured data from image and video has affected their performance, e.g., they are easily influenced by multi-views, occlusion, clothes, and object carrying conditions. This paper addresses these problems using a realistic 3-dimensional (3D) human structural data and sequential pattern learning framework with top-down attention modulating mechanism based on Hierarchical Temporal Memory (HTM). First, an accurate 2-dimensional (2D) to 3D human body pose and shape semantic parameters estimation method is proposed, which exploits the advantages of an instance-level body parsing model and a virtual dressing method. Second, by using gait semantic folding, the estimated body parameters are encoded using a sparse 2D matrix to construct the structural gait semantic image. In order to achieve time-based gait recognition, an HTM Network is constructed to obtain the sequence-level gait sparse distribution representations (SL-GSDRs). A top-down attention mechanism is introduced to deal with various conditions including multi-views by refining the SL-GSDRs, according to prior knowledge. The proposed gait learning model not only aids gait recognition tasks to overcome the difficulties in real application scenarios but also provides the structured gait semantic images for visual cognition. Experimental analyses on CMU MoBo, CASIA B, TUM-IITKGP, and KY4D datasets show a significant performance gain in terms of accuracy and robustness
Gait Recognition Using Period-Based Phase Synchronization for Low Frame-Rate Videos
AbstractβThis paper proposes a method for period-based gait trajectory matching in the eigenspace using phase syn-chronization for low frame-rate videos. First, a gait period is detected by maximizing the normalized autocorrelation of the gait silhouette sequence for the temporal axis. Next, a gait silhouette sequence is expressed as a trajectory in the eigenspace and the gait phase is synchronized by time stretching and time shifting of the trajectory based on the detected period. In addition, multiple period-based matching results are integrated via statistical procedures for more robust matching in the presence of fluctuations among gait sequences. Results of experiments conducted with 185 subjects to evaluate the performance of the gait verification with various spatial and temporal resolutions, demonstrate the effectiveness of the proposed method. Keywords-gait recognition; low frame rate; phase synchro-nization; gait period; PCA I
GaitSet: Regarding Gait as a Set for Cross-View Gait Recognition
As a unique biometric feature that can be recognized at a distance, gait has
broad applications in crime prevention, forensic identification and social
security. To portray a gait, existing gait recognition methods utilize either a
gait template, where temporal information is hard to preserve, or a gait
sequence, which must keep unnecessary sequential constraints and thus loses the
flexibility of gait recognition. In this paper we present a novel perspective,
where a gait is regarded as a set consisting of independent frames. We propose
a new network named GaitSet to learn identity information from the set. Based
on the set perspective, our method is immune to permutation of frames, and can
naturally integrate frames from different videos which have been filmed under
different scenarios, such as diverse viewing angles, different clothes/carrying
conditions. Experiments show that under normal walking conditions, our
single-model method achieves an average rank-1 accuracy of 95.0% on the CASIA-B
gait dataset and an 87.1% accuracy on the OU-MVLP gait dataset. These results
represent new state-of-the-art recognition accuracy. On various complex
scenarios, our model exhibits a significant level of robustness. It achieves
accuracies of 87.2% and 70.4% on CASIA-B under bag-carrying and coat-wearing
walking conditions, respectively. These outperform the existing best methods by
a large margin. The method presented can also achieve a satisfactory accuracy
with a small number of frames in a test sample, e.g., 82.5% on CASIA-B with
only 7 frames. The source code has been released at
https://github.com/AbnerHqC/GaitSet.Comment: AAAI 2019, code is available at https://github.com/AbnerHqC/GaitSe
Human tracking and segmentation supported by silhouette-based gait recognition
Abstract β Gait recognition has recently gained attention as an effective approach to identify individuals at a distance from a camera. Most existing gait recognition algorithms assume that people have been tracked and silhouettes have been segmented successfully. Tacking and segmentation are, however, very difficult especially for articulated objects such as human beings. Therefore, we present an integrated algorithm for tracking and segmentation supported by gait recognition. After the tracking module produces initial results consisting of bounding boxes and foreground likelihood images, the gait recognition module searches for the optimal silhouette-based gait models corresponding to the results. Then, the segmentation module tries to segment people out using the provided gait silhouette sequence as shape priors. Experiments on real video sequences show the effectiveness of the proposed approach. I