96 research outputs found

    MODELING AND ANALYSIS OF WRINKLES ON AGING HUMAN FACES

    Get PDF
    The analysis and modeling of aging human faces has been extensively studied in the past decade. Most of this work is based on matching learning techniques focused on appearance of faces at different ages incorporating facial features such as face shape/geometry and patch-based texture features. However, we do not find much work done on the analysis of facial wrinkles in general and specific to a person. The goal of this dissertation is to analyse and model facial wrinkles for different applications. Facial wrinkles are challenging low-level image features to analyse. In general, skin texture has drastically varying appearance due to its characteristic physical properties. A skin patch looks very different when viewed or illuminated from different angles. This makes subtle skin features like facial wrinkles difficult to be detected in images acquired in uncontrolled imaging settings. In this dissertation, we examine the image properties of wrinkles i.e. intensity gradients and geometric properties and use them for several applications including low-level image processing for automatic detection/localization of wrinkles, soft biometrics and removal of wrinkles using digital inpainting. First, we present results of detection/localization of wrinkles in images using Marked Point Process (MPP). Wrinkles are modeled as sequences of line segments in a Bayesian framework which incorporates a prior probability model based on the likely geometric properties of wrinkles and a data likelihood term based on image intensity gradients. Wrinkles are localized by sampling the posterior probability using a Reversible Jump Markov Chain Monte Carlo (RJMCMC) algorithm. We also present an evaluation algorithm to quantitatively evaluate the detection and false alarm rate of our algorithm and conduct experiments with images taken in uncontrolled settings. The MPP model, despite its promising localization results, requires a large number of iterations in the RJMCMC algorithm to reach global minimum resulting in considerable computation time. This motivated us to adopt a deterministic approach based on image morphology for fast localization of facial wrinkles. We propose image features based on Gabor filter banks to highlight subtle curvilinear discontinuities in skin texture caused by wrinkles. Then, image morphology is used to incorporate geometric constraints to localize curvilinear shapes of wrinkles at image sites of large Gabor filter responses. We conduct experiments on two sets of low and high resolution images to demonstrate faster and visually better localization results as compared to those obtained by MPP modeling. As a next application, we investigate the user-drawn and automatically detected wrinkles as a pattern for their discriminative power as a soft biometrics to recognize subjects from their wrinkle patterns only. A set of facial wrinkles from an image is treated as a curve pattern and used for subject recognition. Given the wrinkle patterns from a query and gallery images, several distance measures are calculated between the two patterns to quantify the similarity between them. This is done by finding the possible correspondences between curves from the two patterns using a simple bipartite graph matching algorithm. Then several metrics are used to calculate the similarity between the two wrinkle patterns. These metrics are based on Hausdorff distance and curve-to-curve correspondences. We conduct experiments on data sets of both hand drawn and automatically detected wrinkles. Finally, we apply digital inpainting to automatically remove wrinkles from facial images. Digital image inpainting refers to filling in the holes of arbitrary shapes in images so that they seem to be part of the original image. The inpainting methods target either the structure or the texture of an image or both. There are two limitations of existing inpainting methods for the removal of wrinkles. First, the differences in the attributes of structure and texture requires different inpainting methods. Facial wrinkles do not fall strictly under the category of structure or texture and can be considered as some where in between. Second, almost all of the image inpainting techniques are supervised i.e. the area/gap to be filled is provided by user interaction and the algorithms attempt to find the suitable image portion automatically. We present an unsupervised image inpainting method where facial regions with wrinkles are detected automatically using their characteristic intensity gradients and removed by painting the regions by the surrounding skin texture

    Modeling of Facial Wrinkles for Applications in Computer Vision

    Get PDF
    International audienceAnalysis and modeling of aging human faces have been extensively studied in the past decade for applications in computer vision such as age estimation, age progression and face recognition across aging. Most of this research work is based on facial appearance and facial features such as face shape, geometry, location of landmarks and patch-based texture features. Despite the recent availability of higher resolution, high quality facial images, we do not find much work on the image analysis of local facial features such as wrinkles specifically. For the most part, modeling of facial skin texture, fine lines and wrinkles has been a focus in computer graphics research for photo-realistic rendering applications. In computer vision, very few aging related applications focus on such facial features. Where several survey papers can be found on facial aging analysis in computer vision, this chapter focuses specifically on the analysis of facial wrinkles in the context of several applications. Facial wrinkles can be categorized as subtle discontinuities or cracks in surrounding inhomogeneous skin texture and pose challenges to being detected/localized in images. First, we review commonly used image features to capture the intensity gradients caused by facial wrinkles and then present research in modeling and analysis of facial wrinkles as aging texture or curvilinear objects for different applications. The reviewed applications include localization or detection of wrinkles in facial images , incorporation of wrinkles for more realistic age progression, analysis for age estimation and inpainting/removal of wrinkles for facial retouching

    Š’ŠµŠ±-Š“Š¾Š“Š°Ń‚Š¾Šŗ Š“Š»Ń рŠ¾Š·ŠæіŠ·Š½Š°Š²Š°Š½Š½Ń Š·Š¼Š¾Ń€ŃˆŠ¾Šŗ Š¾Š±Š»Šøччя

    Get PDF
    Facial recognition technology is named one of the main trends of recent years. Itā€™s wide range of applications, such as access control, biometrics, video surveillance and many other interactive humanmachine systems. Facial landmarks can be described as key characteristics of the human face. Commonly found landmarks are, for example, eyes, nose or mouth corners. Analyzing these key points is useful for a variety of computer vision use cases, including biometrics, face tracking, or emotion detection. Different methods produce different facial landmarks. Some methods use only basic facial landmarks, while others bring out more detail. We use 68 facial markup, which is a common format for many datasets. Cloud computing creates all the necessary conditions for the successful implementation of even the most complex tasks. We created a web application using the Django framework, Python language, OpenCv and Dlib libraries to recognize faces in the image. The purpose of our work is to create a software system for face recognition in the photo and identify wrinkles on the face. The algorithm for determining the presence and location of various types of wrinkles and determining their geometric determination on the face is programmed.Š¢ŠµŃ…Š½Š¾Š»Š¾Š³Ń–ŃŽ рŠ¾Š·ŠæіŠ·Š½Š°Š²Š°Š½Š½Ń Š¾Š±Š»Šøч Š½Š°Š·ŠøŠ²Š°ŃŽŃ‚ŃŒ Š¾Š“Š½ŠøŠ¼ Š· Š³Š¾Š»Š¾Š²Š½Šøх трŠµŠ½Š“іŠ² Š¾ŃŃ‚Š°Š½Š½Ń–Ń… рŠ¾ŠŗіŠ². Š¦Šµ шŠøрŠ¾ŠŗŠøŠ¹ сŠæŠµŠŗтр Š·Š°ŃŃ‚Š¾ŃŃƒŠ²Š°Š½ŃŒ, тŠ°ŠŗŠøх яŠŗ ŠŗŠ¾Š½Ń‚Ń€Š¾Š»ŃŒ Š“Š¾ŃŃ‚ŃƒŠæу, Š±Ń–Š¾Š¼ŠµŃ‚Ń€Ń–Ń, Š²Ń–Š“ŠµŠ¾ŃŠæŠ¾ŃŃ‚ŠµŃ€ŠµŠ¶ŠµŠ½Š½Ń тŠ° Š±Š°Š³Š°Ń‚Š¾ іŠ½ŃˆŠøх іŠ½Ń‚ŠµŃ€Š°ŠŗтŠøŠ²Š½Šøх Š»ŃŽŠ“ŠøŠ½Š¾-Š¼Š°ŃˆŠøŠ½Š½Šøх сŠøстŠµŠ¼. Š›ŠøцьŠ¾Š²Ń– Š¾Ń€Ń–Ń”Š½Ń‚ŠøрŠø Š¼Š¾Š¶Š½Š° Š¾ŠæŠøсŠ°Ń‚Šø яŠŗ ŠŗŠ»ŃŽŃ‡Š¾Š²Ń– хŠ°Ń€Š°ŠŗтŠµŃ€ŠøстŠøŠŗŠø Š»ŃŽŠ“сьŠŗŠ¾Š³Š¾ Š¾Š±Š»Šøччя. ŠŠ°Š¹ŠæŠ¾ŃˆŠøрŠµŠ½Ń–ŃˆŠøŠ¼Šø Š¾Ń€Ń–Ń”Š½Ń‚ŠøрŠ°Š¼Šø є, Š½Š°ŠæрŠøŠŗŠ»Š°Š“, Š¾Ń‡Ń–, ŠŗутŠø Š½Š¾ŃŠ° Š°Š±Š¾ рŠ¾Ń‚Š°. ŠŠ½Š°Š»Ń–Š· цŠøх ŠŗŠ»ŃŽŃ‡Š¾Š²Šøх тŠ¾Ń‡Š¾Šŗ ŠŗŠ¾Ń€ŠøсŠ½ŠøŠ¹ Š“Š»Ń ріŠ·Š½Šøх Š²ŠøŠæŠ°Š“ŠŗіŠ² Š²ŠøŠŗŠ¾Ń€ŠøстŠ°Š½Š½Ń ŠŗŠ¾Š¼Šæ'ютŠµŃ€Š½Š¾Š³Š¾ Š·Š¾Ń€Ńƒ, Š²ŠŗŠ»ŃŽŃ‡Š°ŃŽŃ‡Šø Š±Ń–Š¾Š¼ŠµŃ‚рію, Š²Ń–Š“стŠµŠ¶ŠµŠ½Š½Ń Š¾Š±Š»Šøччя Š°Š±Š¾ Š²ŠøяŠ²Š»ŠµŠ½Š½Ń ŠµŠ¼Š¾Ń†Ń–Š¹. Š Ń–Š·Š½Ń– Š¼ŠµŃ‚Š¾Š“Šø стŠ²Š¾Ń€ŃŽŃŽŃ‚ŃŒ ріŠ·Š½Ń– Š¾Ń€Ń–Ń”Š½Ń‚ŠøрŠø Š¾Š±Š»Šøччя. Š”ŠµŃŠŗі Š¼ŠµŃ‚Š¾Š“Šø Š²ŠøŠŗŠ¾Ń€ŠøстŠ¾Š²ŃƒŃŽŃ‚ŃŒ тіŠ»ŃŒŠŗŠø Š±Š°Š·Š¾Š²Ń– Š¾Ń€Ń–Ń”Š½Ń‚ŠøрŠø Š¾Š±Š»Šøччя, Š² тŠ¾Š¹ чŠ°Ń яŠŗ іŠ½ŃˆŃ– Š²ŠøŠ“іŠ»ŃŃŽŃ‚ŃŒ Š±Ń–Š»ŃŒŃˆŠµ Š“ŠµŃ‚Š°Š»ŠµŠ¹. ŠœŠø Š²ŠøŠŗŠ¾Ń€ŠøстŠ¾Š²ŃƒŃ”Š¼Š¾ 68 рŠ¾Š·Š¼Ń–Ń‚Š¾Šŗ Š¾Š±Š»Šøччя, яŠŗі є Š·Š°Š³Š°Š»ŃŒŠ½ŠøŠ¼ фŠ¾Ń€Š¼Š°Ń‚Š¾Š¼ Š“Š»Ń Š±Š°Š³Š°Ń‚ŃŒŠ¾Ń… Š½Š°Š±Š¾Ń€Ń–Š² Š“Š°Š½Šøх. Š„Š¼Š°Ń€Š½Ń– Š¾Š±Ń‡ŠøсŠ»ŠµŠ½Š½Ń стŠ²Š¾Ń€ŃŽŃŽŃ‚ŃŒ Š²ŃŃ– Š½ŠµŠ¾Š±Ń…Ń–Š“Š½Ń– уŠ¼Š¾Š²Šø Š“Š»Ń усŠæішŠ½Š¾Ń— рŠµŠ°Š»Ń–Š·Š°Ń†Ń–Ń— Š½Š°Š²Ń–Ń‚ŃŒ Š½Š°Š¹ŃŠŗŠ»Š°Š“Š½Ń–ŃˆŠøх Š·Š°Š²Š“Š°Š½ŃŒ. ŠœŠø стŠ²Š¾Ń€ŠøŠ»Šø Š²ŠµŠ±-Š“Š¾Š“Š°Ń‚Š¾Šŗ Š· Š²ŠøŠŗŠ¾Ń€ŠøстŠ°Š½Š½ŃŠ¼ фрŠµŠ¹Š¼Š²Š¾Ń€Šŗу Django, Š¼Š¾Š²Šø Python, Š±Ń–Š±Š»Ń–Š¾Ń‚ŠµŠŗ OpenCv тŠ° Dlib Š“Š»Ń рŠ¾Š·ŠæіŠ·Š½Š°Š²Š°Š½Š½Ń Š¾Š±Š»Šøч Š½Š° Š·Š¾Š±Ń€Š°Š¶ŠµŠ½Š½Ń–. ŠœŠµŃ‚Š¾ŃŽ Š½Š°ŃˆŠ¾Ń— рŠ¾Š±Š¾Ń‚Šø є стŠ²Š¾Ń€ŠµŠ½Š½Ń ŠæрŠ¾Š³Ń€Š°Š¼Š½Š¾Š³Š¾ ŠŗŠ¾Š¼ŠæŠ»ŠµŠŗсу Š“Š»Ń рŠ¾Š·ŠæіŠ·Š½Š°Š²Š°Š½Š½Ń Š¾Š±Š»Šøччя Š½Š° фŠ¾Ń‚Š¾Š³Ń€Š°Ń„Ń–Ń— тŠ° Š²ŠøяŠ²Š»ŠµŠ½Š½Ń Š·Š¼Š¾Ń€ŃˆŠ¾Šŗ Š½Š° Š¾Š±Š»Šøччі. Š—Š°ŠæрŠ¾Š³Ń€Š°Š¼Š¾Š²Š°Š½Š¾ Š°Š»Š³Š¾Ń€ŠøтŠ¼ Š²ŠøŠ·Š½Š°Ń‡ŠµŠ½Š½Ń Š½Š°ŃŠ²Š½Š¾ŃŃ‚Ń– тŠ° рŠ¾Š·Ń‚Š°ŃˆŃƒŠ²Š°Š½Š½Ń ріŠ·Š½Šøх тŠøŠæіŠ² Š·Š¼Š¾Ń€ŃˆŠ¾Šŗ тŠ° Š²ŠøŠ·Š½Š°Ń‡ŠµŠ½Š½Ń їх Š³ŠµŠ¾Š¼ŠµŃ‚Ń€ŠøчŠ½Š¾Ń— Š²ŠøŠ·Š½Š°Ń‡ŠµŠ½Š¾ŃŃ‚Ń– Š½Š° Š¾Š±Š»Šøччі

    Face age estimation using wrinkle patterns

    Get PDF
    Face age estimation is a challenging problem due to the variation of craniofacial growth, skin texture, gender and race. With recent growth in face age estimation research, wrinkles received attention from a number of research, as it is generally perceived as aging feature and soft biometric for person identification. In a face image, wrinkle is a discontinuous and arbitrary line pattern that varies in different face regions and subjects. Existing wrinkle detection algorithms and wrinkle-based features are not robust for face age estimation. They are either weakly represented or not validated against the ground truth. The primary aim of this thesis is to develop a robust wrinkle detection method and construct novel wrinkle-based methods for face age estimation. First, Hybrid Hessian Filter (HHF) is proposed to segment the wrinkles using the directional gradient and a ridge-valley Gaussian kernel. Second, Hessian Line Tracking (HLT) is proposed for wrinkle detection by exploring the wrinkle connectivity of surrounding pixels using a cross-sectional profile. Experimental results showed that HLT outperforms other wrinkle detection algorithms with an accuracy of 84% and 79% on the datasets of FORERUS and FORERET while HHF achieves 77% and 49%, respectively. Third, Multi-scale Wrinkle Patterns (MWP) is proposed as a novel feature representation for face age estimation using the wrinkle location, intensity and density. Fourth, Hybrid Aging Patterns (HAP) is proposed as a hybrid pattern for face age estimation using Facial Appearance Model (FAM) and MWP. Fifth, Multi-layer Age Regression (MAR) is proposed as a hierarchical model in complementary of FAM and MWP for face age estimation. For performance assessment of age estimation, four datasets namely FGNET, MORPH, FERET and PAL with different age ranges and sample sizes are used as benchmarks. Results showed that MAR achieves the lowest Mean Absolute Error (MAE) of 3.00 ( 4.14) on FERET and HAP scores a comparable MAE of 3.02 ( 2.92) as state of the art. In conclusion, wrinkles are important features and the uniqueness of this pattern should be considered in developing a robust model for face age estimation

    A Survey on Ear Biometrics

    No full text
    Recognizing people by their ear has recently received significant attention in the literature. Several reasons account for this trend: first, ear recognition does not suffer from some problems associated with other non contact biometrics, such as face recognition; second, it is the most promising candidate for combination with the face in the context of multi-pose face recognition; and third, the ear can be used for human recognition in surveillance videos where the face may be occluded completely or in part. Further, the ear appears to degrade little with age. Even though, current ear detection and recognition systems have reached a certain level of maturity, their success is limited to controlled indoor conditions. In addition to variation in illumination, other open research problems include hair occlusion; earprint forensics; ear symmetry; ear classification; and ear individuality. This paper provides a detailed survey of research conducted in ear detection and recognition. It provides an up-to-date review of the existing literature revealing the current state-of-art for not only those who are working in this area but also for those who might exploit this new approach. Furthermore, it offers insights into some unsolved ear recognition problems as well as ear databases available for researchers

    Automatic analysis of facial actions: a survey

    Get PDF
    As one of the most comprehensive and objective ways to describe facial expressions, the Facial Action Coding System (FACS) has recently received significant attention. Over the past 30 years, extensive research has been conducted by psychologists and neuroscientists on various aspects of facial expression analysis using FACS. Automating FACS coding would make this research faster and more widely applicable, opening up new avenues to understanding how we communicate through facial expressions. Such an automated process can also potentially increase the reliability, precision and temporal resolution of coding. This paper provides a comprehensive survey of research into machine analysis of facial actions. We systematically review all components of such systems: pre-processing, feature extraction and machine coding of facial actions. In addition, the existing FACS-coded facial expression databases are summarised. Finally, challenges that have to be addressed to make automatic facial action analysis applicable in real-life situations are extensively discussed. There are two underlying motivations for us to write this survey paper: the first is to provide an up-to-date review of the existing literature, and the second is to offer some insights into the future of machine recognition of facial actions: what are the challenges and opportunities that researchers in the field face

    Timing is everything: A spatio-temporal approach to the analysis of facial actions

    No full text
    This thesis presents a fully automatic facial expression analysis system based on the Facial Action Coding System (FACS). FACS is the best known and the most commonly used system to describe facial activity in terms of facial muscle actions (i.e., action units, AUs). We will present our research on the analysis of the morphological, spatio-temporal and behavioural aspects of facial expressions. In contrast with most other researchers in the field who use appearance based techniques, we use a geometric feature based approach. We will argue that that approach is more suitable for analysing facial expression temporal dynamics. Our system is capable of explicitly exploring the temporal aspects of facial expressions from an input colour video in terms of their onset (start), apex (peak) and offset (end). The fully automatic system presented here detects 20 facial points in the first frame and tracks them throughout the video. From the tracked points we compute geometry-based features which serve as the input to the remainder of our systems. The AU activation detection system uses GentleBoost feature selection and a Support Vector Machine (SVM) classifier to find which AUs were present in an expression. Temporal dynamics of active AUs are recognised by a hybrid GentleBoost-SVM-Hidden Markov model classifier. The system is capable of analysing 23 out of 27 existing AUs with high accuracy. The main contributions of the work presented in this thesis are the following: we have created a method for fully automatic AU analysis with state-of-the-art recognition results. We have proposed for the first time a method for recognition of the four temporal phases of an AU. We have build the largest comprehensive database of facial expressions to date. We also present for the first time in the literature two studies for automatic distinction between posed and spontaneous expressions

    Unifying the Visible and Passive Infrared Bands: Homogeneous and Heterogeneous Multi-Spectral Face Recognition

    Get PDF
    Face biometrics leverages tools and technology in order to automate the identification of individuals. In most cases, biometric face recognition (FR) can be used for forensic purposes, but there remains the issue related to the integration of technology into the legal system of the court. The biggest challenge with the acceptance of the face as a modality used in court is the reliability of such systems under varying pose, illumination and expression, which has been an active and widely explored area of research over the last few decades (e.g. same-spectrum or homogeneous matching). The heterogeneous FR problem, which deals with matching face images from different sensors, should be examined for the benefit of military and law enforcement applications as well. In this work we are concerned primarily with visible band images (380-750 nm) and the infrared (IR) spectrum, which has become an area of growing interest.;For homogeneous FR systems, we formulate and develop an efficient, semi-automated, direct matching-based FR framework, that is designed to operate efficiently when face data is captured using either visible or passive IR sensors. Thus, it can be applied in both daytime and nighttime environments. First, input face images are geometrically normalized using our pre-processing pipeline prior to feature-extraction. Then, face-based features including wrinkles, veins, as well as edges of facial characteristics, are detected and extracted for each operational band (visible, MWIR, and LWIR). Finally, global and local face-based matching is applied, before fusion is performed at the score level. Although this proposed matcher performs well when same-spectrum FR is performed, regardless of spectrum, a challenge exists when cross-spectral FR matching is performed. The second framework is for the heterogeneous FR problem, and deals with the issue of bridging the gap across the visible and passive infrared (MWIR and LWIR) spectrums. Specifically, we investigate the benefits and limitations of using synthesized visible face images from thermal and vice versa, in cross-spectral face recognition systems when utilizing canonical correlation analysis (CCA) and locally linear embedding (LLE), a manifold learning technique for dimensionality reduction. Finally, by conducting an extensive experimental study we establish that the combination of the proposed synthesis and demographic filtering scheme increases system performance in terms of rank-1 identification rate

    Gaussian processes for modeling of facial expressions

    Get PDF
    Automated analysis of facial expressions has been gaining significant attention over the past years. This stems from the fact that it constitutes the primal step toward developing some of the next-generation computer technologies that can make an impact in many domains, ranging from medical imaging and health assessment to marketing and education. No matter the target application, the need to deploy systems under demanding, real-world conditions that can generalize well across the population is urgent. Hence, careful consideration of numerous factors has to be taken prior to designing such a system. The work presented in this thesis focuses on tackling two important problems in automated analysis of facial expressions: (i) view-invariant facial expression analysis; (ii) modeling of the structural patterns in the face, in terms of well coordinated facial muscle movements. Driven by the necessity for efficient and accurate inference mechanisms we explore machine learning techniques based on the probabilistic framework of Gaussian processes (GPs). Our ultimate goal is to design powerful models that can efficiently handle imagery with spontaneously displayed facial expressions, and explain in detail the complex configurations behind the human face in real-world situations. To effectively decouple the head pose and expression in the presence of large out-of-plane head rotations we introduce a manifold learning approach based on multi-view learning strategies. Contrary to the majority of existing methods that typically treat the numerous poses as individual problems, in this model we first learn a discriminative manifold shared by multiple views of a facial expression. Subsequently, we perform facial expression classification in the expression manifold. Hence, the pose normalization problem is solved by aligning the facial expressions from different poses in a common latent space. We demonstrate that the recovered manifold can efficiently generalize to various poses and expressions even from a small amount of training data, while also being largely robust to corrupted image features due to illumination variations. State-of-the-art performance is achieved in the task of facial expression classification of basic emotions. The methods that we propose for learning the structure in the configuration of the muscle movements represent some of the first attempts in the field of analysis and intensity estimation of facial expressions. In these models, we extend our multi-view approach to exploit relationships not only in the input features but also in the multi-output labels. The structure of the outputs is imposed into the recovered manifold either from heuristically defined hard constraints, or in an auto-encoded manner, where the structure is learned automatically from the input data. The resulting models are proven to be robust to data with imbalanced expression categories, due to our proposed Bayesian learning of the target manifold. We also propose a novel regression approach based on product of GP experts where we take into account people's individual expressiveness in order to adapt the learned models on each subject. We demonstrate the superior performance of our proposed models on the task of facial expression recognition and intensity estimation.Open Acces
    • ā€¦
    corecore