257 research outputs found

    Occluded iris classification and segmentation using self-customized artificial intelligence models and iterative randomized Hough transform

    Get PDF
    A fast and accurate iris recognition system is presented for noisy iris images, mainly the noises due to eye occlusion and from specular reflection. The proposed recognition system will adopt a self-customized support vector machine (SVM) and convolution neural network (CNN) classification models, where the models are built according to the iris texture GLCM and automated deep features datasets that are extracted exclusively from each subject individually. The image processing techniques used were optimized, whether the processing of iris region segmentation using iterative randomized Hough transform (IRHT), or the processing of the classification, where few significant features are considered, based on singular value decomposition (SVD) analysis, for testing the moving window matrix class if it is iris or non-iris. The iris segments matching techniques are optimized by extracting, first, the largest parallel-axis rectangle inscribed in the classified occluded-iris binary image, where its corresponding iris region is crosscorrelated with the same subject’s iris reference image for obtaining the most correlated iris segments in the two eye images. Finally, calculating the iriscode Hamming distance of the two most correlated segments to identify the subject’s unique iris pattern with high accuracy, security, and reliability

    Feature extraction using two dimensional (2D) legendre wavelet filter for partial iris recognition

    Get PDF
    An increasing need for biometrics recognition systems has grown substantially to address the issues of recognition and identification, especially in highly dense areas such as airports, train stations, and financial transactions. Evidence of these can be seen in some airports and also the implementation of these technologies in our mobile phones. Among the most popular biometric technologies include facial, fingerprints, and iris recognition. The iris recognition is considered by many researchers to be the most accurate and reliable form of biometric recognition because iris can neither be surgically operated with a chance of losing slight nor change due to aging. However, presently most iris recognition systems available can only recognize iris image with frontal-looking and high-quality images. Angular image and partially capture image cannot be authenticated with the existing method of iris recognition. This research investigates the possibility of developing a technique for recognition partially captured iris image. The technique is designed to process the iris image at 50%, 25%, 16.5%, and 12.5% and to find a threshold for a minimum amount of iris region required to authenticate the individual. The research also developed and implemented two Dimensional (2D) Legendre wavelet filter for the iris feature extraction. The Legendre wavelet filter is to enhance the feature extraction technique. Selected iris images from CASIA, UBIRIS, and MMU database were used to test the accuracy of the introduced technique. The technique was able to produce recognition accuracy between 70 – 90% CASIA-interval with 92.25% accuracy, CASIA-distance with 86.25%, UBIRIS with 74.95%, and MMU with 94.45%

    Human-Centric Machine Vision

    Get PDF
    Recently, the algorithms for the processing of the visual information have greatly evolved, providing efficient and effective solutions to cope with the variability and the complexity of real-world environments. These achievements yield to the development of Machine Vision systems that overcome the typical industrial applications, where the environments are controlled and the tasks are very specific, towards the use of innovative solutions to face with everyday needs of people. The Human-Centric Machine Vision can help to solve the problems raised by the needs of our society, e.g. security and safety, health care, medical imaging, and human machine interface. In such applications it is necessary to handle changing, unpredictable and complex situations, and to take care of the presence of humans

    Homogeneous and Heterogeneous Face Recognition: Enhancing, Encoding and Matching for Practical Applications

    Get PDF
    Face Recognition is the automatic processing of face images with the purpose to recognize individuals. Recognition task becomes especially challenging in surveillance applications, where images are acquired from a long range in the presence of difficult environments. Short Wave Infrared (SWIR) is an emerging imaging modality that is able to produce clear long range images in difficult environments or during night time. Despite the benefits of the SWIR technology, matching SWIR images against a gallery of visible images presents a challenge, since the photometric properties of the images in the two spectral bands are highly distinct.;In this dissertation, we describe a cross spectral matching method that encodes magnitude and phase of multi-spectral face images filtered with a bank of Gabor filters. The magnitude of filtered images is encoded with Simplified Weber Local Descriptor (SWLD) and Local Binary Pattern (LBP) operators. The phase is encoded with Generalized Local Binary Pattern (GLBP) operator. Encoded multi-spectral images are mapped into a histogram representation and cross matched by applying symmetric Kullback-Leibler distance. Performance of the developed algorithm is demonstrated on TINDERS database that contains long range SWIR and color images acquired at a distance of 2, 50, and 106 meters.;Apart from long acquisition range, other variations and distortions such as pose variation, motion and out of focus blur, and uneven illumination may be observed in multispectral face images. Recognition performance of the face recognition matcher can be greatly affected by these distortions. It is important, therefore, to ensure that matching is performed on high quality images. Poor quality images have to be either enhanced or discarded. This dissertation addresses the problem of selecting good quality samples.;The last chapters of the dissertation suggest a number of modifications applied to the cross spectral matching algorithm for matching low resolution color images in near-real time. We show that the method that encodes the magnitude of Gabor filtered images with the SWLD operator guarantees high recognition rates. The modified method (Gabor-SWLD) is adopted in a camera network set up where cameras acquire several views of the same individual. The designed algorithm and software are fully automated and optimized to perform recognition in near-real time. We evaluate the recognition performance and the processing time of the method on a small dataset collected at WVU

    Pattern Recognition

    Get PDF
    Pattern recognition is a very wide research field. It involves factors as diverse as sensors, feature extraction, pattern classification, decision fusion, applications and others. The signals processed are commonly one, two or three dimensional, the processing is done in real- time or takes hours and days, some systems look for one narrow object class, others search huge databases for entries with at least a small amount of similarity. No single person can claim expertise across the whole field, which develops rapidly, updates its paradigms and comprehends several philosophical approaches. This book reflects this diversity by presenting a selection of recent developments within the area of pattern recognition and related fields. It covers theoretical advances in classification and feature extraction as well as application-oriented works. Authors of these 25 works present and advocate recent achievements of their research related to the field of pattern recognition

    Recognition of Nonideal Iris Images Using Shape Guided Approach and Game Theory

    Get PDF
    Most state-of-the-art iris recognition algorithms claim to perform with a very high recognition accuracy in a strictly controlled environment. However, their recognition accuracies significantly decrease when the acquired images are affected by different noise factors including motion blur, camera diffusion, head movement, gaze direction, camera angle, reflections, contrast, luminosity, eyelid and eyelash occlusions, and problems due to contraction and dilation. The main objective of this thesis is to develop a nonideal iris recognition system by using active contour methods, Genetic Algorithms (GAs), shape guided model, Adaptive Asymmetrical Support Vector Machines (AASVMs) and Game Theory (GT). In this thesis, the proposed iris recognition method is divided into two phases: (1) cooperative iris recognition, and (2) noncooperative iris recognition. While most state-of-the-art iris recognition algorithms have focused on the preprocessing of iris images, recently, important new directions have been identified in iris biometrics research. These include optimal feature selection and iris pattern classification. In the first phase, we propose an iris recognition scheme based on GAs and asymmetrical SVMs. Instead of using the whole iris region, we elicit the iris information between the collarette and the pupil boundary to suppress the effects of eyelid and eyelash occlusions and to minimize the matching error. In the second phase, we process the nonideal iris images that are captured in unconstrained situations and those affected by several nonideal factors. The proposed noncooperative iris recognition method is further divided into three approaches. In the first approach of the second phase, we apply active contour-based curve evolution approaches to segment the inner/outer boundaries accurately from the nonideal iris images. The proposed active contour-based approaches show a reasonable performance when the iris/sclera boundary is separated by a blurred boundary. In the second approach, we describe a new iris segmentation scheme using GT to elicit iris/pupil boundary from a nonideal iris image. We apply a parallel game-theoretic decision making procedure by modifying Chakraborty and Duncan's algorithm to form a unified approach, which is robust to noise and poor localization and less affected by weak iris/sclera boundary. Finally, to further improve the segmentation performance, we propose a variational model to localize the iris region belonging to the given shape space using active contour method, a geometric shape prior and the Mumford-Shah functional. The verification and identification performance of the proposed scheme is validated using four challenging nonideal iris datasets, namely, the ICE 2005, the UBIRIS Version 1, the CASIA Version 3 Interval, and the WVU Nonideal, plus the non-homogeneous combined dataset. We have conducted several sets of experiments and finally, the proposed approach has achieved a Genuine Accept Rate (GAR) of 97.34% on the combined dataset at the fixed False Accept Rate (FAR) of 0.001% with an Equal Error Rate (EER) of 0.81%. The highest Correct Recognition Rate (CRR) obtained by the proposed iris recognition system is 97.39%

    Computer Vision Based Early Intraocular Pressure Assessment From Frontal Eye Images

    Get PDF
    Intraocular Pressure (IOP) in general, refers to the pressure in the eyes. Gradual increase of IOP and high IOP are conditions or symptoms that may lead to certain diseases such as glaucoma, and therefore, must be closely monitored. While the pressure in the eye increases, different parts of the eye may become affected until the eye parts are damaged. An effective way to prevent rise in eye pressure is by early detection. Exiting IOP monitoring tools include eye tests at clinical facilities and computer-aided techniques from fundus and optic nerves images. In this work, a new computer vision-based smart healthcare framework is presented to evaluate the intraocular pressure risk from frontal eye images early-on. The framework determines the status of IOP by analyzing frontal eye images using image processing and machine learning techniques. A database of images from the Princess Basma Hospital was used in this work. The database contains 400 eye images; 200 images with normal IOP and 200 high eye pressure case images. This study proposes novel features for IOP determination from two experiments. The first experiment extracts the sclera using circular hough transform, after which four features are extracted from the whole sclera. These features are mean redness level, red area percentage, contour area and contour height. The pupil/iris diameter ratio feature is also extracted from the frontal eye image after a series of pre-processing techniques. The second experiment extracts the sclera and iris segment using a fully conventional neural network technique, after which six features are extracted from only part of the segmented sclera and iris. The features include mean redness level, red area percentage, contour area, contour distance and contour angle along with the pupil/iris diameter ratio. Once the features are extracted, classification techniques are applied in order to train and test the images and features to obtain the status of the patients in terms of eye pressure. For the first experiment, neural network and support vector machine algorithms were adopted in order to detect the status of intraocular pressure. The second experiment adopted support vector machine and decision tree algorithms to detect the status of intraocular pressure. For both experiments, the framework detects the status of IOP (normal or high IOP) with high accuracies. This computer vison-based approach produces evidence of the relationship between the extracted frontal eye image features and IOP, which has not been previously investigated through automated image processing and machine learning techniques from frontal eye images
    • …
    corecore