815 research outputs found
Signal processing and machine learning techniques for human verification based on finger textures
PhD ThesisIn recent years, Finger Textures (FTs) have attracted considerable
attention as potential biometric characteristics. They can provide
robust recognition performance as they have various human-speci c
features, such as wrinkles and apparent lines distributed along the
inner surface of all ngers. The main topic of this thesis is verifying
people according to their unique FT patterns by exploiting signal
processing and machine learning techniques.
A Robust Finger Segmentation (RFS) method is rst proposed to
isolate nger images from a hand area. It is able to detect the ngers
as objects from a hand image. An e cient adaptive nger
segmentation method is also suggested to address the problem of
alignment variations in the hand image called the Adaptive and Robust
Finger Segmentation (ARFS) method.
A new Multi-scale Sobel Angles Local Binary Pattern (MSALBP)
feature extraction method is proposed which combines the Sobel
direction angles with the Multi-Scale Local Binary Pattern (MSLBP).
Moreover, an enhanced method called the Enhanced Local Line Binary
Pattern (ELLBP) is designed to e ciently analyse the FT patterns. As
a result, a powerful human veri cation scheme based on nger Feature
Level Fusion with a Probabilistic Neural Network (FLFPNN) is
proposed. A multi-object fusion method, termed the Finger
Contribution Fusion Neural Network (FCFNN), combines the
contribution scores of the nger objects.
The veri cation performances are examined in the case of missing FT
areas. Consequently, to overcome nger regions which are poorly
imaged a method is suggested to salvage missing FT elements by
exploiting the information embedded within the trained Probabilistic
Neural Network (PNN). Finally, a novel method to produce a Receiver
Operating Characteristic (ROC) curve from a PNN is suggested.
Furthermore, additional development to this method is applied to
generate the ROC graph from the FCFNN.
Three databases are employed for evaluation: The Hong Kong
Polytechnic University Contact-free 3D/2D (PolyU3D2D), Indian
Institute of Technology (IIT) Delhi and Spectral 460nm (S460) from
the CASIA Multi-Spectral (CASIAMS) databases. Comparative
simulation studies con rm the e ciency of the proposed methods for
human veri cation.
The main advantage of both segmentation approaches, the RFS and
ARFS, is that they can collect all the FT features. The best results
have been benchmarked for the ELLBP feature extraction with the
FCFNN, where the best Equal Error Rate (EER) values for the three
databases PolyU3D2D, IIT Delhi and CASIAMS (S460) have been
achieved 0.11%, 1.35% and 0%, respectively. The proposed salvage
approach for the missing feature elements has the capability to enhance
the veri cation performance for the FLFPNN. Moreover, ROC graphs
have been successively established from the PNN and FCFNN.the ministry of higher
education and scientific research in Iraq (MOHESR); the Technical
college of Mosul; the Iraqi Cultural Attach e; the active people in the
MOHESR, who strongly supported Iraqi students
Design and Real-World Application of Novel Machine Learning Techniques for Improving Face Recognition Algorithms
Recent progress in machine learning has made possible the development of real-world face recognition applications that can match face images as good as or better than humans. However, several challenges remain unsolved. In this PhD thesis, some of these challenges are studied and novel machine learning techniques to improve the performance of real-world face recognition applications are proposed.
Current face recognition algorithms based on deep learning techniques are able to achieve outstanding accuracy when dealing with face images taken in unconstrained environments. However, training these algorithms is often costly due to the very large datasets and the high computational resources needed. On the other hand, traditional methods for face recognition are better suited when these requirements cannot be satisfied. This PhD thesis presents new techniques for both traditional and deep learning methods. In particular, a novel traditional face recognition method that combines texture and shape features together with subspace representation techniques is first presented. The proposed method is lightweight and can be trained quickly with small datasets. This method is used for matching face images scanned from identity documents against face images stored in the biometric chip of such documents. Next, two new techniques to increase the performance of face recognition methods based on convolutional neural networks are presented. Specifically, a novel training strategy that increases face recognition accuracy when dealing with face images presenting occlusions, and a new loss function that improves the performance of the triplet loss function are proposed. Finally, the problem of collecting large face datasets is considered, and a novel method based on generative adversarial networks to synthesize both face images of existing subjects in a dataset and face images of new subjects is proposed. The accuracy of existing face recognition algorithms can be increased by training with datasets augmented with the synthetic face images generated by the proposed method. In addition to the main contributions, this thesis provides a comprehensive literature review of face recognition methods and their evolution over the years.
A significant amount of the work presented in this PhD thesis is the outcome of a 3-year-long research project partially funded by Innovate UK as part of a Knowledge Transfer Partnership between University of Hertfordshire and IDscan Biometrics Ltd (partnership number: 009547)
Toward Open-Set Face Recognition
Much research has been conducted on both face identification and face
verification, with greater focus on the latter. Research on face identification
has mostly focused on using closed-set protocols, which assume that all probe
images used in evaluation contain identities of subjects that are enrolled in
the gallery. Real systems, however, where only a fraction of probe sample
identities are enrolled in the gallery, cannot make this closed-set assumption.
Instead, they must assume an open set of probe samples and be able to
reject/ignore those that correspond to unknown identities. In this paper, we
address the widespread misconception that thresholding verification-like scores
is a good way to solve the open-set face identification problem, by formulating
an open-set face identification protocol and evaluating different strategies
for assessing similarity. Our open-set identification protocol is based on the
canonical labeled faces in the wild (LFW) dataset. Additionally to the known
identities, we introduce the concepts of known unknowns (known, but
uninteresting persons) and unknown unknowns (people never seen before) to the
biometric community. We compare three algorithms for assessing similarity in a
deep feature space under an open-set protocol: thresholded verification-like
scores, linear discriminant analysis (LDA) scores, and an extreme value machine
(EVM) probabilities. Our findings suggest that thresholding EVM probabilities,
which are open-set by design, outperforms thresholding verification-like
scores.Comment: Accepted for Publication in CVPR 2017 Biometrics Worksho
Recommended from our members
Evaluation and analysis of hybrid intelligent pattern recognition techniques for speaker identification
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The rapid momentum of the technology progress in the recent years has led to a tremendous rise in the use of biometric authentication systems. The objective of this research is to investigate the problem
of identifying a speaker from its voice regardless of the content (i.e.
text-independent), and to design efficient methods of combining face and voice in producing a robust authentication system.
A novel approach towards speaker identification is developed using
wavelet analysis, and multiple neural networks including Probabilistic
Neural Network (PNN), General Regressive Neural Network (GRNN)and Radial Basis Function-Neural Network (RBF NN) with the AND
voting scheme. This approach is tested on GRID and VidTIMIT cor-pora and comprehensive test results have been validated with state-
of-the-art approaches. The system was found to be competitive and it improved the recognition rate by 15% as compared to the classical Mel-frequency Cepstral Coe±cients (MFCC), and reduced the recognition time by 40% compared to Back Propagation Neural Network (BPNN), Gaussian Mixture Models (GMM) and Principal Component Analysis (PCA).
Another novel approach using vowel formant analysis is implemented using Linear Discriminant Analysis (LDA). Vowel formant based speaker identification is best suitable for real-time implementation and requires only a few bytes of information to be stored for each speaker, making it both storage and time efficient. Tested on GRID and Vid-TIMIT, the proposed scheme was found to be 85.05% accurate when Linear Predictive Coding (LPC) is used to extract the vowel formants, which is much higher than the accuracy of BPNN and GMM. Since the proposed scheme does not require any training time other than creating a small database of vowel formants, it is faster as well. Furthermore, an increasing number of speakers makes it di±cult for BPNN and GMM to sustain their accuracy, but the proposed score-based methodology stays almost linear.
Finally, a novel audio-visual fusion based identification system is implemented using GMM and MFCC for speaker identi¯cation and PCA for face recognition. The results of speaker identification and face recognition are fused at different levels, namely the feature, score and decision levels. Both the score-level and decision-level (with OR voting) fusions were shown to outperform the feature-level fusion in terms of accuracy and error resilience. The result is in line with the distinct nature of the two modalities which lose themselves when combined at the feature-level. The GRID and VidTIMIT test results validate that
the proposed scheme is one of the best candidates for the fusion of
face and voice due to its low computational time and high recognition accuracy
Deep finger texture learning for verifying people
Finger Texture (FT) is currently attracting significant attentions in the area of human recognition. Finger texture covers the area between the lower knuckle of the finger and the upper phalanx before the fingerprint. It involves rich features which can be efficiently used as a biometric characteristic. In this paper, we contribute to this growing area by proposing a new verification
approach, i.e., Deep Finger Texture Learning (DFTL). To the best of our knowledge, this is the first time that deep learning is employed for recognizing people by using the FT characteristic. Four databases have been used to evaluate the proposed method: the Hong Kong Polytechnic University Contact-free 3D/2D (PolyU2D), Indian Institute of Technology Delhi (IITD), CASIA Blue spectral (CASIA-BLU) corresponding to spectral 460nm and CASIA White spectral (CASIA-WHT) from the CASIA Multi-Spectral images database. The obtained results have shown superior performance compared with recent literature. The Verification Accuracies (VAs) have attained 100%, 98.65%, 100% and 98% for the four databases of PolyU2D, IITD, CASIA-BLU and CASIA-WHT, respectively
- …