15 research outputs found
A study of the temporal relationship between eye actions and facial expressions
A dissertation submitted in ful llment of the requirements for the
degree of Master of Science
in the
School of Computer Science and Applied Mathematics
Faculty of Science
August 15, 2017Facial expression recognition is one of the most common means of communication used
for complementing spoken word. However, people have grown to master ways of ex-
hibiting deceptive expressions. Hence, it is imperative to understand di erences in
expressions mostly for security purposes among others. Traditional methods employ
machine learning techniques in di erentiating real and fake expressions. However, this
approach does not always work as human subjects can easily mimic real expressions with
a bit of practice. This study presents an approach that evaluates the time related dis-
tance that exists between eye actions and an exhibited expression. The approach gives
insights on some of the most fundamental characteristics of expressions. The study fo-
cuses on nding and understanding the temporal relationship that exists between eye
blinks and smiles. It further looks at the relationship that exits between eye closure and
pain expressions. The study incorporates active appearance models (AAM) for feature
extraction and support vector machines (SVM) for classi cation. It tests extreme learn-
ing machines (ELM) in both smile and pain studies, which in turn, attains excellent
results than predominant algorithms like the SVM. The study shows that eye blinks
are highly correlated with the beginning of a smile in posed smiles while eye blinks are
highly correlated with the end of a smile in spontaneous smiles. A high correlation is
observed between eye closure and pain in spontaneous pain expressions. Furthermore,
this study brings about ideas that lead to potential applications such as lie detection
systems, robust health care monitoring systems and enhanced animation design systems
among others.MT 201
A study of the temporal relationship between eye actions and facial expressions
A dissertation submitted in ful llment of the requirements for the
degree of Master of Science
in the
School of Computer Science and Applied Mathematics
Faculty of Science
August 15, 2017Facial expression recognition is one of the most common means of communication used
for complementing spoken word. However, people have grown to master ways of ex-
hibiting deceptive expressions. Hence, it is imperative to understand di erences in
expressions mostly for security purposes among others. Traditional methods employ
machine learning techniques in di erentiating real and fake expressions. However, this
approach does not always work as human subjects can easily mimic real expressions with
a bit of practice. This study presents an approach that evaluates the time related dis-
tance that exists between eye actions and an exhibited expression. The approach gives
insights on some of the most fundamental characteristics of expressions. The study fo-
cuses on nding and understanding the temporal relationship that exists between eye
blinks and smiles. It further looks at the relationship that exits between eye closure and
pain expressions. The study incorporates active appearance models (AAM) for feature
extraction and support vector machines (SVM) for classi cation. It tests extreme learn-
ing machines (ELM) in both smile and pain studies, which in turn, attains excellent
results than predominant algorithms like the SVM. The study shows that eye blinks
are highly correlated with the beginning of a smile in posed smiles while eye blinks are
highly correlated with the end of a smile in spontaneous smiles. A high correlation is
observed between eye closure and pain in spontaneous pain expressions. Furthermore,
this study brings about ideas that lead to potential applications such as lie detection
systems, robust health care monitoring systems and enhanced animation design systems
among others.MT 201
Machine Analysis of Facial Expressions
No abstract
Timing is everything: A spatio-temporal approach to the analysis of facial actions
This thesis presents a fully automatic facial expression analysis system based on the Facial Action
Coding System (FACS). FACS is the best known and the most commonly used system to describe
facial activity in terms of facial muscle actions (i.e., action units, AUs). We will present our research
on the analysis of the morphological, spatio-temporal and behavioural aspects of facial expressions.
In contrast with most other researchers in the field who use appearance based techniques, we use a
geometric feature based approach. We will argue that that approach is more suitable for analysing
facial expression temporal dynamics. Our system is capable of explicitly exploring the temporal
aspects of facial expressions from an input colour video in terms of their onset (start), apex (peak)
and offset (end).
The fully automatic system presented here detects 20 facial points in the first frame and tracks them
throughout the video. From the tracked points we compute geometry-based features which serve as
the input to the remainder of our systems. The AU activation detection system uses GentleBoost
feature selection and a Support Vector Machine (SVM) classifier to find which AUs were present in an
expression. Temporal dynamics of active AUs are recognised by a hybrid GentleBoost-SVM-Hidden
Markov model classifier. The system is capable of analysing 23 out of 27 existing AUs with high
accuracy.
The main contributions of the work presented in this thesis are the following: we have created a
method for fully automatic AU analysis with state-of-the-art recognition results. We have proposed
for the first time a method for recognition of the four temporal phases of an AU. We have build the
largest comprehensive database of facial expressions to date. We also present for the first time in the
literature two studies for automatic distinction between posed and spontaneous expressions
Recommended from our members
The computational face for facial emotion analysis: Computer based emotion analysis from the face
Facial expressions are considered to be the most revealing way of understanding the human psychological state during face-to-face communication. It is believed that a more natural interaction between humans and machines can be undertaken through the detailed understanding of the different facial expressions which imitate the manner by which humans communicate with each other.
In this research, we study the different aspects of facial emotion detection, analysis and investigate possible hidden identity clues within the facial expressions. We study a deeper aspect of facial expressions whereby we try to identify gender and human identity - which can be considered as a form of emotional biometric - using only the dynamic characteristics of the smile expressions. Further, we present a statistical model for analysing the relationship between facial features and Duchenne (real) and non-Duchenne (posed) smiles. Thus, we identify that the expressions in the eyes contain discriminating features between Duchenne and non-Duchenne smiles.
Our results indicate that facial expressions can be identified through facial movement analysis models where we get an accuracy rate of 86% for classifying the six universal facial expressions and 94% for classifying the common 18 facial action units. Further, we successfully identify the gender using only the dynamic characteristics of the smile expression whereby we obtain an 86% classification rate. Likewise, we present a framework to study the possibility of using the smile as a biometric whereby we show that the human smile is unique and stable.Al-Zaytoonah Universit
Facial Micro- and Macro-Expression Spotting and Generation Methods
Facial micro-expression (ME) recognition requires face movement interval as input, but computer methods in spotting ME are still underperformed. This is due to lacking large-scale long video dataset and ME generation methods are in their infancy. This thesis presents methods to address data deficiency issues and introduces a new method for spotting macro- and micro-expressions simultaneously.
This thesis introduces SAMM Long Videos (SAMM-LV), which contains 147 annotated long videos, and develops a baseline method to facilitate ME Grand Challenge 2020. Further, a reference-guided style transfer of StarGANv2 is experimented on SAMM-LV to generate a synthetic dataset, namely SAMM-SYNTH. The quality of SAMM-SYNTH is evaluated by using facial action units detected by OpenFace. Quantitative measurement shows high correlations on two Action Units (AU12 and AU6) of the original and synthetic data.
In facial expression spotting, a two-stream 3D-Convolutional Neural Network with temporal oriented frame skips that can spot micro- and macro-expression simultaneously is proposed. This method achieves state-of-the-art performance in SAMM-LV and is competitive in CAS(ME)2, it was used as the baseline result of ME Grand Challenge 2021. The F1-score improves to 0.1036 when trained with composite data consisting of SAMM-LV and SAMMSYNTH. On the unseen ME Grand Challenge 2022 evaluation dataset, it achieves F1-score of 0.1531.
Finally, a new sequence generation method to explore the capability of deep learning network is proposed. It generates spontaneous facial expressions by using only two input sequences without any labels. SSIM and NIQE were used for image quality analysis and the generated data achieved 0.87 and 23.14. By visualising the movements using optical flow value and absolute frame differences, this method demonstrates its potential in generating subtle ME. For realism evaluation, the generated videos were rated by using two facial expression recognition networks
Artificial Intelligence Tools for Facial Expression Analysis.
Inner emotions show visibly upon the human face and are understood as a basic guide to an individual’s inner world. It is, therefore, possible to determine a person’s attitudes and the effects of others’ behaviour on their deeper feelings through examining facial expressions. In real world applications, machines that interact with people need strong facial expression recognition. This recognition is seen to hold advantages for varied applications in affective computing, advanced human-computer interaction, security, stress and depression analysis, robotic systems, and machine learning. This thesis starts by proposing a benchmark of dynamic versus static methods for facial Action Unit (AU) detection. AU activation is a set of local individual facial muscle parts that occur in unison constituting a natural facial expression event. Detecting AUs automatically can provide explicit benefits since it considers both static and dynamic facial features. For this research, AU occurrence activation detection was conducted by extracting features (static and dynamic) of both nominal hand-crafted and deep learning representation from each static image of a video. This confirmed the superior ability of a pretrained model that leaps in performance. Next, temporal modelling was investigated to detect the underlying temporal variation phases using supervised and unsupervised methods from dynamic sequences. During these processes, the importance of stacking dynamic on top of static was discovered in encoding deep features for learning temporal information when combining the spatial and temporal schemes simultaneously. Also, this study found that fusing both temporal and temporal features will give more long term temporal pattern information. Moreover, we hypothesised that using an unsupervised method would enable the leaching of invariant information from dynamic textures. Recently, fresh cutting-edge developments have been created by approaches based on Generative Adversarial Networks (GANs). In the second section of this thesis, we propose a model based on the adoption of an unsupervised DCGAN for the facial features’ extraction and classification to achieve the following: the creation of facial expression images under different arbitrary poses (frontal, multi-view, and in the wild), and the recognition of emotion categories and AUs, in an attempt to resolve the problem of recognising the static seven classes of emotion in the wild. Thorough experimentation with the proposed cross-database performance demonstrates that this approach can improve the generalization results. Additionally, we showed that the features learnt by the DCGAN process are poorly suited to encoding facial expressions when observed under multiple views, or when trained from a limited number of positive examples. Finally, this research focuses on disentangling identity from expression for facial expression recognition. A novel technique was implemented for emotion recognition from a single monocular image. A large-scale dataset (Face vid) was created from facial image videos which were rich in variations and distribution of facial dynamics, appearance, identities, expressions, and 3D poses. This dataset was used to train a DCNN (ResNet) to regress the expression parameters from a 3D Morphable Model jointly with a back-end classifier
A new descriptor for smile classification based on cascade classifier in unconstrained scenarios
In the development of human–machine interfaces, facial expression analysis has attracted
considerable attention, as it provides a natural and efficient way of communication. Congruence
between facial and behavioral inference in face processing is considered a serious challenge that needs
to be solved in the near future. Automatic facial expression is a difficult classification issue because
of the high interclass variability caused by the significant interdependence of the environmental
conditions on the face appearance caused by head pose, scale, and illumination occlusions from their
variances. In this paper, an adaptive model for smile classification is suggested that integrates a
row-transform-based feature extraction algorithm and a cascade classifier to increase the precision
of facial recognition. We suggest a histogram-based cascade smile classification method utilizing different facial features. The candidate feature set was designed based on the first order histogram probability, and a cascade classifier with a variety of parameters was used at the classification stage.
Row transformation is used to exclude any unnecessary coefficients in a vector, thereby enhancing the discriminatory capacity of the extracted features and reducing the sophistication of the calculations. Cascading gives the opportunity to train an extremely precise classification by taking a weighted average of poor learners’ decisions. Through accumulating positive and negative images of a single object, this algorithm can build a complete classifier capable of classifying different smiles in a limited amount of time (near real time) and with a high level of precision (92.2–98.8%) as opposed to other algorithms by large margins (5% compared with traditional neural network and 2% compared with Deep Neural Network based methods)
Automatic analysis of facial actions: a survey
As one of the most comprehensive and objective ways to describe facial expressions, the Facial Action Coding System (FACS) has recently received significant attention. Over the past 30 years, extensive research has been conducted by psychologists and neuroscientists on various aspects of facial expression analysis using FACS. Automating FACS coding would make this research faster and more widely applicable, opening up new avenues to understanding how we communicate through facial expressions. Such an automated process can also potentially increase the reliability, precision and temporal resolution of coding. This paper provides a comprehensive survey of research into machine analysis of facial actions. We systematically review all components of such systems: pre-processing, feature extraction and machine coding of facial actions. In addition, the existing FACS-coded facial expression databases are summarised. Finally, challenges that have to be addressed to make automatic facial action analysis applicable in real-life situations are extensively discussed. There are two underlying motivations for us to write this survey paper: the first is to provide an up-to-date review of the existing literature, and the second is to offer some insights into the future of machine recognition of facial actions: what are the challenges and opportunities that researchers in the field face