3,318 research outputs found

    Multi-scale Deep Learning Architectures for Person Re-identification

    Full text link
    Person Re-identification (re-id) aims to match people across non-overlapping camera views in a public space. It is a challenging problem because many people captured in surveillance videos wear similar clothes. Consequently, the differences in their appearance are often subtle and only detectable at the right location and scales. Existing re-id models, particularly the recently proposed deep learning based ones match people at a single scale. In contrast, in this paper, a novel multi-scale deep learning model is proposed. Our model is able to learn deep discriminative feature representations at different scales and automatically determine the most suitable scales for matching. The importance of different spatial locations for extracting discriminative features is also learned explicitly. Experiments are carried out to demonstrate that the proposed model outperforms the state-of-the art on a number of benchmarksComment: 9 pages, 3 figures, accepted by ICCV 201

    Curvelet and Ridgelet-based Multimodal Biometric Recognition System using Weighted Similarity Approach

    Get PDF
    Biometric security artifacts for establishing the identity of a person with high confidence have evoked enormous interest in security and access control applications for the past few years. Biometric systems based solely on unimodal biometrics often suffer from problems such as noise, intra-class variations and spoof attacks. This paper presents a novel multimodal biometric recognition system by integrating three biometric traits namely iris, fingerprint and face using weighted similarity approach. In this work, the multi-resolution features are extracted independently from query images using curvelet and ridgelet transforms, and are then compared to the enrolled templates stored in the database containing features of each biometric trait. The final decision is made by normalizing the feature vectors, assigning different weights to the modalities and fusing the computed scores using score combination techniques. This system is tested with the public unimodal databases such as CASIA–Iris-V3-Interval, FVC2004, ORL and self-built multimodal databases. Experimental results obtained shows that the designed system achieves an excellent recognition rate of 98.75 per cent and 100 per cent for the public and self-built databases respectively and provides ultra high security than unimodal biometric systems.Defence Science Journal, 2014, 64(2), pp. 106-114. DOI: http://dx.doi.org/10.14429/dsj.64.346

    Infrared face recognition: a comprehensive review of methodologies and databases

    Full text link
    Automatic face recognition is an area with immense practical potential which includes a wide range of commercial and law enforcement applications. Hence it is unsurprising that it continues to be one of the most active research areas of computer vision. Even after over three decades of intense research, the state-of-the-art in face recognition continues to improve, benefitting from advances in a range of different research fields such as image processing, pattern recognition, computer graphics, and physiology. Systems based on visible spectrum images, the most researched face recognition modality, have reached a significant level of maturity with some practical success. However, they continue to face challenges in the presence of illumination, pose and expression changes, as well as facial disguises, all of which can significantly decrease recognition accuracy. Amongst various approaches which have been proposed in an attempt to overcome these limitations, the use of infrared (IR) imaging has emerged as a particularly promising research direction. This paper presents a comprehensive and timely review of the literature on this subject. Our key contributions are: (i) a summary of the inherent properties of infrared imaging which makes this modality promising in the context of face recognition, (ii) a systematic review of the most influential approaches, with a focus on emerging common trends as well as key differences between alternative methodologies, (iii) a description of the main databases of infrared facial images available to the researcher, and lastly (iv) a discussion of the most promising avenues for future research.Comment: Pattern Recognition, 2014. arXiv admin note: substantial text overlap with arXiv:1306.160

    Cross-Modality High-Frequency Transformer for MR Image Super-Resolution

    Full text link
    Improving the resolution of magnetic resonance (MR) image data is critical to computer-aided diagnosis and brain function analysis. Higher resolution helps to capture more detailed content, but typically induces to lower signal-to-noise ratio and longer scanning time. To this end, MR image super-resolution has become a widely-interested topic in recent times. Existing works establish extensive deep models with the conventional architectures based on convolutional neural networks (CNN). In this work, to further advance this research field, we make an early effort to build a Transformer-based MR image super-resolution framework, with careful designs on exploring valuable domain prior knowledge. Specifically, we consider two-fold domain priors including the high-frequency structure prior and the inter-modality context prior, and establish a novel Transformer architecture, called Cross-modality high-frequency Transformer (Cohf-T), to introduce such priors into super-resolving the low-resolution (LR) MR images. Comprehensive experiments on two datasets indicate that Cohf-T achieves new state-of-the-art performance
    • …
    corecore